diff --git a/lib/include/__clang_cuda_math.h b/lib/include/__clang_cuda_math.h index acb26ad345..538556f394 100644 --- a/lib/include/__clang_cuda_math.h +++ b/lib/include/__clang_cuda_math.h @@ -166,6 +166,8 @@ __DEVICE__ long long llrint(double __a) { return __nv_llrint(__a); } __DEVICE__ long long llrintf(float __a) { return __nv_llrintf(__a); } __DEVICE__ long long llround(double __a) { return __nv_llround(__a); } __DEVICE__ long long llroundf(float __a) { return __nv_llroundf(__a); } +__DEVICE__ double round(double __a) { return __nv_round(__a); } +__DEVICE__ float roundf(float __a) { return __nv_roundf(__a); } __DEVICE__ double log(double __a) { return __nv_log(__a); } __DEVICE__ double log10(double __a) { return __nv_log10(__a); } __DEVICE__ float log10f(float __a) { return __nv_log10f(__a); } @@ -270,8 +272,6 @@ __DEVICE__ float rnorm4df(float __a, float __b, float __c, float __d) { __DEVICE__ float rnormf(int __dim, const float *__t) { return __nv_rnormf(__dim, __t); } -__DEVICE__ double round(double __a) { return __nv_round(__a); } -__DEVICE__ float roundf(float __a) { return __nv_roundf(__a); } __DEVICE__ double rsqrt(double __a) { return __nv_rsqrt(__a); } __DEVICE__ float rsqrtf(float __a) { return __nv_rsqrtf(__a); } __DEVICE__ double scalbn(double __a, int __b) { return __nv_scalbn(__a, __b); } diff --git a/lib/include/__clang_cuda_runtime_wrapper.h b/lib/include/__clang_cuda_runtime_wrapper.h index f88c39a9b6..f401964bd5 100644 --- a/lib/include/__clang_cuda_runtime_wrapper.h +++ b/lib/include/__clang_cuda_runtime_wrapper.h @@ -349,9 +349,14 @@ extern "C" { __device__ int vprintf(const char *, const char *); __device__ void free(void *) __attribute((nothrow)); __device__ void *malloc(size_t) __attribute((nothrow)) __attribute__((malloc)); + +// __assertfail() used to have a `noreturn` attribute. Unfortunately that +// contributed to triggering the longstanding bug in ptxas when assert was used +// in sufficiently convoluted code. See +// https://bugs.llvm.org/show_bug.cgi?id=27738 for the details. __device__ void __assertfail(const char *__message, const char *__file, unsigned __line, const char *__function, - size_t __charSize) __attribute__((noreturn)); + size_t __charSize); // In order for standard assert() macro on linux to work we need to // provide device-side __assert_fail() diff --git a/lib/include/__clang_hip_cmath.h b/lib/include/__clang_hip_cmath.h index cd22a2df95..7342705434 100644 --- a/lib/include/__clang_hip_cmath.h +++ b/lib/include/__clang_hip_cmath.h @@ -14,6 +14,7 @@ #error "This file is for HIP and OpenMP AMDGCN device compilation only." #endif +#if !defined(__HIPCC_RTC__) #if defined(__cplusplus) #include #include @@ -21,6 +22,7 @@ #endif #include #include +#endif // !defined(__HIPCC_RTC__) #pragma push_macro("__DEVICE__") #define __DEVICE__ static __device__ inline __attribute__((always_inline)) @@ -34,6 +36,9 @@ __DEVICE__ long abs(long __n) { return ::labs(__n); } __DEVICE__ float fma(float __x, float __y, float __z) { return ::fmaf(__x, __y, __z); } +#if !defined(__HIPCC_RTC__) +// The value returned by fpclassify is platform dependent, therefore it is not +// supported by hipRTC. __DEVICE__ int fpclassify(float __x) { return __builtin_fpclassify(FP_NAN, FP_INFINITE, FP_NORMAL, FP_SUBNORMAL, FP_ZERO, __x); @@ -42,11 +47,51 @@ __DEVICE__ int fpclassify(double __x) { return __builtin_fpclassify(FP_NAN, FP_INFINITE, FP_NORMAL, FP_SUBNORMAL, FP_ZERO, __x); } +#endif // !defined(__HIPCC_RTC__) + __DEVICE__ float frexp(float __arg, int *__exp) { return ::frexpf(__arg, __exp); } + +#if defined(__OPENMP_AMDGCN__) +// For OpenMP we work around some old system headers that have non-conforming +// `isinf(float)` and `isnan(float)` implementations that return an `int`. We do +// this by providing two versions of these functions, differing only in the +// return type. To avoid conflicting definitions we disable implicit base +// function generation. That means we will end up with two specializations, one +// per type, but only one has a base function defined by the system header. +#pragma omp begin declare variant match( \ + implementation = {extension(disable_implicit_base)}) + +// FIXME: We lack an extension to customize the mangling of the variants, e.g., +// add a suffix. This means we would clash with the names of the variants +// (note that we do not create implicit base functions here). To avoid +// this clash we add a new trait to some of them that is always true +// (this is LLVM after all ;)). It will only influence the mangled name +// of the variants inside the inner region and avoid the clash. +#pragma omp begin declare variant match(implementation = {vendor(llvm)}) + +__DEVICE__ int isinf(float __x) { return ::__isinff(__x); } +__DEVICE__ int isinf(double __x) { return ::__isinf(__x); } +__DEVICE__ int isfinite(float __x) { return ::__finitef(__x); } +__DEVICE__ int isfinite(double __x) { return ::__finite(__x); } +__DEVICE__ int isnan(float __x) { return ::__isnanf(__x); } +__DEVICE__ int isnan(double __x) { return ::__isnan(__x); } + +#pragma omp end declare variant +#endif // defined(__OPENMP_AMDGCN__) + +__DEVICE__ bool isinf(float __x) { return ::__isinff(__x); } +__DEVICE__ bool isinf(double __x) { return ::__isinf(__x); } __DEVICE__ bool isfinite(float __x) { return ::__finitef(__x); } __DEVICE__ bool isfinite(double __x) { return ::__finite(__x); } +__DEVICE__ bool isnan(float __x) { return ::__isnanf(__x); } +__DEVICE__ bool isnan(double __x) { return ::__isnan(__x); } + +#if defined(__OPENMP_AMDGCN__) +#pragma omp end declare variant +#endif // defined(__OPENMP_AMDGCN__) + __DEVICE__ bool isgreater(float __x, float __y) { return __builtin_isgreater(__x, __y); } @@ -59,8 +104,6 @@ __DEVICE__ bool isgreaterequal(float __x, float __y) { __DEVICE__ bool isgreaterequal(double __x, double __y) { return __builtin_isgreaterequal(__x, __y); } -__DEVICE__ bool isinf(float __x) { return ::__isinff(__x); } -__DEVICE__ bool isinf(double __x) { return ::__isinf(__x); } __DEVICE__ bool isless(float __x, float __y) { return __builtin_isless(__x, __y); } @@ -79,8 +122,6 @@ __DEVICE__ bool islessgreater(float __x, float __y) { __DEVICE__ bool islessgreater(double __x, double __y) { return __builtin_islessgreater(__x, __y); } -__DEVICE__ bool isnan(float __x) { return ::__isnanf(__x); } -__DEVICE__ bool isnan(double __x) { return ::__isnan(__x); } __DEVICE__ bool isnormal(float __x) { return __builtin_isnormal(__x); } __DEVICE__ bool isnormal(double __x) { return __builtin_isnormal(__x); } __DEVICE__ bool isunordered(float __x, float __y) { @@ -207,11 +248,117 @@ template struct __hip_enable_if {}; template struct __hip_enable_if { typedef __T type; }; +namespace __hip { +template struct is_integral { + enum { value = 0 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; + +// ToDo: specializes is_arithmetic<_Float16> +template struct is_arithmetic { + enum { value = 0 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; + +struct true_type { + static const __constant__ bool value = true; +}; +struct false_type { + static const __constant__ bool value = false; +}; + +template struct is_same : public false_type {}; +template struct is_same<__T, __T> : public true_type {}; + +template struct add_rvalue_reference { typedef __T &&type; }; + +template typename add_rvalue_reference<__T>::type declval(); + // decltype is only available in C++11 and above. #if __cplusplus >= 201103L // __hip_promote -namespace __hip { - template struct __numeric_type { static void __test(...); static _Float16 __test(_Float16); @@ -227,8 +374,8 @@ template struct __numeric_type { // No support for long double, use double instead. static double __test(long double); - typedef decltype(__test(std::declval<_Tp>())) type; - static const bool value = !std::is_same::value; + typedef decltype(__test(declval<_Tp>())) type; + static const bool value = !is_same::value; }; template <> struct __numeric_type { static const bool value = true; }; @@ -271,18 +418,17 @@ public: template class __promote : public __promote_imp<_A1, _A2, _A3> {}; - -} // namespace __hip #endif //__cplusplus >= 201103L +} // namespace __hip // __HIP_OVERLOAD1 is used to resolve function calls with integer argument to // avoid compilation error due to ambibuity. e.g. floor(5) is resolved with // floor(double). #define __HIP_OVERLOAD1(__retty, __fn) \ template \ - __DEVICE__ typename __hip_enable_if::is_integer, \ - __retty>::type \ - __fn(__T __x) { \ + __DEVICE__ \ + typename __hip_enable_if<__hip::is_integral<__T>::value, __retty>::type \ + __fn(__T __x) { \ return ::__fn((double)__x); \ } @@ -293,8 +439,7 @@ class __promote : public __promote_imp<_A1, _A2, _A3> {}; #define __HIP_OVERLOAD2(__retty, __fn) \ template \ __DEVICE__ typename __hip_enable_if< \ - std::numeric_limits<__T1>::is_specialized && \ - std::numeric_limits<__T2>::is_specialized, \ + __hip::is_arithmetic<__T1>::value && __hip::is_arithmetic<__T2>::value, \ typename __hip::__promote<__T1, __T2>::type>::type \ __fn(__T1 __x, __T2 __y) { \ typedef typename __hip::__promote<__T1, __T2>::type __result_type; \ @@ -303,16 +448,14 @@ class __promote : public __promote_imp<_A1, _A2, _A3> {}; #else #define __HIP_OVERLOAD2(__retty, __fn) \ template \ - __DEVICE__ \ - typename __hip_enable_if::is_specialized && \ - std::numeric_limits<__T2>::is_specialized, \ - __retty>::type \ - __fn(__T1 __x, __T2 __y) { \ + __DEVICE__ typename __hip_enable_if<__hip::is_arithmetic<__T1>::value && \ + __hip::is_arithmetic<__T2>::value, \ + __retty>::type \ + __fn(__T1 __x, __T2 __y) { \ return __fn((double)__x, (double)__y); \ } #endif -__HIP_OVERLOAD1(double, abs) __HIP_OVERLOAD1(double, acos) __HIP_OVERLOAD1(double, acosh) __HIP_OVERLOAD1(double, asin) @@ -336,7 +479,9 @@ __HIP_OVERLOAD1(double, floor) __HIP_OVERLOAD2(double, fmax) __HIP_OVERLOAD2(double, fmin) __HIP_OVERLOAD2(double, fmod) +#if !defined(__HIPCC_RTC__) __HIP_OVERLOAD1(int, fpclassify) +#endif // !defined(__HIPCC_RTC__) __HIP_OVERLOAD2(double, hypot) __HIP_OVERLOAD1(int, ilogb) __HIP_OVERLOAD1(bool, isfinite) @@ -382,9 +527,8 @@ __HIP_OVERLOAD2(double, min) #if __cplusplus >= 201103L template __DEVICE__ typename __hip_enable_if< - std::numeric_limits<__T1>::is_specialized && - std::numeric_limits<__T2>::is_specialized && - std::numeric_limits<__T3>::is_specialized, + __hip::is_arithmetic<__T1>::value && __hip::is_arithmetic<__T2>::value && + __hip::is_arithmetic<__T3>::value, typename __hip::__promote<__T1, __T2, __T3>::type>::type fma(__T1 __x, __T2 __y, __T3 __z) { typedef typename __hip::__promote<__T1, __T2, __T3>::type __result_type; @@ -392,33 +536,32 @@ fma(__T1 __x, __T2 __y, __T3 __z) { } #else template -__DEVICE__ - typename __hip_enable_if::is_specialized && - std::numeric_limits<__T2>::is_specialized && - std::numeric_limits<__T3>::is_specialized, - double>::type - fma(__T1 __x, __T2 __y, __T3 __z) { +__DEVICE__ typename __hip_enable_if<__hip::is_arithmetic<__T1>::value && + __hip::is_arithmetic<__T2>::value && + __hip::is_arithmetic<__T3>::value, + double>::type +fma(__T1 __x, __T2 __y, __T3 __z) { return ::fma((double)__x, (double)__y, (double)__z); } #endif template __DEVICE__ - typename __hip_enable_if::is_integer, double>::type + typename __hip_enable_if<__hip::is_integral<__T>::value, double>::type frexp(__T __x, int *__exp) { return ::frexp((double)__x, __exp); } template __DEVICE__ - typename __hip_enable_if::is_integer, double>::type + typename __hip_enable_if<__hip::is_integral<__T>::value, double>::type ldexp(__T __x, int __exp) { return ::ldexp((double)__x, __exp); } template __DEVICE__ - typename __hip_enable_if::is_integer, double>::type + typename __hip_enable_if<__hip::is_integral<__T>::value, double>::type modf(__T __x, double *__exp) { return ::modf((double)__x, __exp); } @@ -426,8 +569,8 @@ __DEVICE__ #if __cplusplus >= 201103L template __DEVICE__ - typename __hip_enable_if::is_specialized && - std::numeric_limits<__T2>::is_specialized, + typename __hip_enable_if<__hip::is_arithmetic<__T1>::value && + __hip::is_arithmetic<__T2>::value, typename __hip::__promote<__T1, __T2>::type>::type remquo(__T1 __x, __T2 __y, int *__quo) { typedef typename __hip::__promote<__T1, __T2>::type __result_type; @@ -435,25 +578,24 @@ __DEVICE__ } #else template -__DEVICE__ - typename __hip_enable_if::is_specialized && - std::numeric_limits<__T2>::is_specialized, - double>::type - remquo(__T1 __x, __T2 __y, int *__quo) { +__DEVICE__ typename __hip_enable_if<__hip::is_arithmetic<__T1>::value && + __hip::is_arithmetic<__T2>::value, + double>::type +remquo(__T1 __x, __T2 __y, int *__quo) { return ::remquo((double)__x, (double)__y, __quo); } #endif template __DEVICE__ - typename __hip_enable_if::is_integer, double>::type + typename __hip_enable_if<__hip::is_integral<__T>::value, double>::type scalbln(__T __x, long int __exp) { return ::scalbln((double)__x, __exp); } template __DEVICE__ - typename __hip_enable_if::is_integer, double>::type + typename __hip_enable_if<__hip::is_integral<__T>::value, double>::type scalbn(__T __x, int __exp) { return ::scalbn((double)__x, __exp); } @@ -468,14 +610,15 @@ __DEVICE__ #endif // defined(__cplusplus) // Define these overloads inside the namespace our standard library uses. +#if !defined(__HIPCC_RTC__) #ifdef _LIBCPP_BEGIN_NAMESPACE_STD _LIBCPP_BEGIN_NAMESPACE_STD #else namespace std { #ifdef _GLIBCXX_BEGIN_NAMESPACE_VERSION _GLIBCXX_BEGIN_NAMESPACE_VERSION -#endif -#endif +#endif // _GLIBCXX_BEGIN_NAMESPACE_VERSION +#endif // _LIBCPP_BEGIN_NAMESPACE_STD // Pull the new overloads we defined above into namespace std. // using ::abs; - This may be considered for C++. @@ -620,11 +763,13 @@ _LIBCPP_END_NAMESPACE_STD #else #ifdef _GLIBCXX_BEGIN_NAMESPACE_VERSION _GLIBCXX_END_NAMESPACE_VERSION -#endif +#endif // _GLIBCXX_BEGIN_NAMESPACE_VERSION } // namespace std -#endif +#endif // _LIBCPP_END_NAMESPACE_STD +#endif // !defined(__HIPCC_RTC__) // Define device-side math functions from on MSVC. +#if !defined(__HIPCC_RTC__) #if defined(_MSC_VER) // Before VS2019, `` is also included in `` and other headers. @@ -658,6 +803,7 @@ __DEVICE__ __attribute__((overloadable)) float _FSinh(float x, float y) { } #endif // defined(__cplusplus) #endif // defined(_MSC_VER) +#endif // !defined(__HIPCC_RTC__) #pragma pop_macro("__DEVICE__") diff --git a/lib/include/__clang_hip_libdevice_declares.h b/lib/include/__clang_hip_libdevice_declares.h index ac98907ad5..8be848ba2a 100644 --- a/lib/include/__clang_hip_libdevice_declares.h +++ b/lib/include/__clang_hip_libdevice_declares.h @@ -138,14 +138,22 @@ __device__ __attribute__((const)) float __ocml_fma_rtn_f32(float, float, float); __device__ __attribute__((const)) float __ocml_fma_rtp_f32(float, float, float); __device__ __attribute__((const)) float __ocml_fma_rtz_f32(float, float, float); -__device__ __attribute__((const)) float -__llvm_amdgcn_cos_f32(float) __asm("llvm.amdgcn.cos.f32"); -__device__ __attribute__((const)) float -__llvm_amdgcn_rcp_f32(float) __asm("llvm.amdgcn.rcp.f32"); -__device__ __attribute__((const)) float -__llvm_amdgcn_rsq_f32(float) __asm("llvm.amdgcn.rsq.f32"); -__device__ __attribute__((const)) float -__llvm_amdgcn_sin_f32(float) __asm("llvm.amdgcn.sin.f32"); +__device__ inline __attribute__((const)) float +__llvm_amdgcn_cos_f32(float __x) { + return __builtin_amdgcn_cosf(__x); +} +__device__ inline __attribute__((const)) float +__llvm_amdgcn_rcp_f32(float __x) { + return __builtin_amdgcn_rcpf(__x); +} +__device__ inline __attribute__((const)) float +__llvm_amdgcn_rsq_f32(float __x) { + return __builtin_amdgcn_rsqf(__x); +} +__device__ inline __attribute__((const)) float +__llvm_amdgcn_sin_f32(float __x) { + return __builtin_amdgcn_sinf(__x); +} // END INTRINSICS // END FLOAT @@ -269,10 +277,14 @@ __device__ __attribute__((const)) double __ocml_fma_rtp_f64(double, double, __device__ __attribute__((const)) double __ocml_fma_rtz_f64(double, double, double); -__device__ __attribute__((const)) double -__llvm_amdgcn_rcp_f64(double) __asm("llvm.amdgcn.rcp.f64"); -__device__ __attribute__((const)) double -__llvm_amdgcn_rsq_f64(double) __asm("llvm.amdgcn.rsq.f64"); +__device__ inline __attribute__((const)) double +__llvm_amdgcn_rcp_f64(double __x) { + return __builtin_amdgcn_rcp(__x); +} +__device__ inline __attribute__((const)) double +__llvm_amdgcn_rsq_f64(double __x) { + return __builtin_amdgcn_rsq(__x); +} __device__ __attribute__((const)) _Float16 __ocml_ceil_f16(_Float16); __device__ _Float16 __ocml_cos_f16(_Float16); diff --git a/lib/include/__clang_hip_math.h b/lib/include/__clang_hip_math.h index 14d91c66b3..1f0982d92e 100644 --- a/lib/include/__clang_hip_math.h +++ b/lib/include/__clang_hip_math.h @@ -13,11 +13,13 @@ #error "This file is for HIP and OpenMP AMDGCN device compilation only." #endif +#if !defined(__HIPCC_RTC__) #if defined(__cplusplus) #include #endif #include #include +#endif // __HIPCC_RTC__ #pragma push_macro("__DEVICE__") #define __DEVICE__ static __device__ inline __attribute__((always_inline)) @@ -36,7 +38,7 @@ template struct __compare_result{}; template<> struct __compare_result { - static const bool valid; + static const __device__ bool valid; }; __DEVICE__ @@ -1260,6 +1262,7 @@ float min(float __x, float __y) { return fminf(__x, __y); } __DEVICE__ double min(double __x, double __y) { return fmin(__x, __y); } +#if !defined(__HIPCC_RTC__) __host__ inline static int min(int __arg1, int __arg2) { return std::min(__arg1, __arg2); } @@ -1267,6 +1270,7 @@ __host__ inline static int min(int __arg1, int __arg2) { __host__ inline static int max(int __arg1, int __arg2) { return std::max(__arg1, __arg2); } +#endif // __HIPCC_RTC__ #endif #pragma pop_macro("__DEVICE__") diff --git a/lib/include/__clang_hip_runtime_wrapper.h b/lib/include/__clang_hip_runtime_wrapper.h index 81a16a265a..73021d256c 100644 --- a/lib/include/__clang_hip_runtime_wrapper.h +++ b/lib/include/__clang_hip_runtime_wrapper.h @@ -18,52 +18,107 @@ #if __HIP__ -#include -#include -#include - #define __host__ __attribute__((host)) #define __device__ __attribute__((device)) #define __global__ __attribute__((global)) #define __shared__ __attribute__((shared)) #define __constant__ __attribute__((constant)) +#define __managed__ __attribute__((managed)) #if !defined(__cplusplus) || __cplusplus < 201103L #define nullptr NULL; #endif +#ifdef __cplusplus +extern "C" { + __attribute__((__visibility__("default"))) + __attribute__((weak)) + __attribute__((noreturn)) + __device__ void __cxa_pure_virtual(void) { + __builtin_trap(); + } + __attribute__((__visibility__("default"))) + __attribute__((weak)) + __attribute__((noreturn)) + __device__ void __cxa_deleted_virtual(void) { + __builtin_trap(); + } +} +#endif //__cplusplus + +#if !defined(__HIPCC_RTC__) +#include +#include +#include +#else +typedef __SIZE_TYPE__ size_t; +// Define macros which are needed to declare HIP device API's without standard +// C/C++ headers. This is for readability so that these API's can be written +// the same way as non-hipRTC use case. These macros need to be popped so that +// they do not pollute users' name space. +#pragma push_macro("NULL") +#pragma push_macro("uint32_t") +#pragma push_macro("uint64_t") +#pragma push_macro("CHAR_BIT") +#pragma push_macro("INT_MAX") +#define NULL (void *)0 +#define uint32_t __UINT32_TYPE__ +#define uint64_t __UINT64_TYPE__ +#define CHAR_BIT __CHAR_BIT__ +#define INT_MAX __INTMAX_MAX__ +#endif // __HIPCC_RTC__ + +typedef __SIZE_TYPE__ __hip_size_t; + +#ifdef __cplusplus +extern "C" { +#endif //__cplusplus + #if __HIP_ENABLE_DEVICE_MALLOC__ -extern "C" __device__ void *__hip_malloc(size_t __size); -extern "C" __device__ void *__hip_free(void *__ptr); -static inline __device__ void *malloc(size_t __size) { +__device__ void *__hip_malloc(__hip_size_t __size); +__device__ void *__hip_free(void *__ptr); +__attribute__((weak)) inline __device__ void *malloc(__hip_size_t __size) { return __hip_malloc(__size); } -static inline __device__ void *free(void *__ptr) { return __hip_free(__ptr); } -#else -static inline __device__ void *malloc(size_t __size) { - __builtin_trap(); - return nullptr; +__attribute__((weak)) inline __device__ void *free(void *__ptr) { + return __hip_free(__ptr); } -static inline __device__ void *free(void *__ptr) { +#else +__attribute__((weak)) inline __device__ void *malloc(__hip_size_t __size) { __builtin_trap(); - return nullptr; + return (void *)0; +} +__attribute__((weak)) inline __device__ void *free(void *__ptr) { + __builtin_trap(); + return (void *)0; } #endif +#ifdef __cplusplus +} // extern "C" +#endif //__cplusplus + #include <__clang_hip_libdevice_declares.h> #include <__clang_hip_math.h> -#if !_OPENMP || __HIP_ENABLE_CUDA_WRAPPER_FOR_OPENMP__ +#if defined(__HIPCC_RTC__) +#include <__clang_hip_cmath.h> +#else #include <__clang_cuda_math_forward_declares.h> #include <__clang_hip_cmath.h> #include <__clang_cuda_complex_builtins.h> - #include #include #include -#endif // !_OPENMP || __HIP_ENABLE_CUDA_WRAPPER_FOR_OPENMP__ +#endif // __HIPCC_RTC__ #define __CLANG_HIP_RUNTIME_WRAPPER_INCLUDED__ 1 - +#if defined(__HIPCC_RTC__) +#pragma pop_macro("NULL") +#pragma pop_macro("uint32_t") +#pragma pop_macro("uint64_t") +#pragma pop_macro("CHAR_BIT") +#pragma pop_macro("INT_MAX") +#endif // __HIPCC_RTC__ #endif // __HIP__ #endif // __CLANG_HIP_RUNTIME_WRAPPER_H__ diff --git a/lib/include/altivec.h b/lib/include/altivec.h index 4d50d47d51..0dd8c85936 100644 --- a/lib/include/altivec.h +++ b/lib/include/altivec.h @@ -41,9 +41,7 @@ #define __ATTRS_o_ai __attribute__((__overloadable__, __always_inline__)) -#ifdef __POWER9_VECTOR__ #include -#endif static __inline__ vector signed char __ATTRS_o_ai vec_perm( vector signed char __a, vector signed char __b, vector unsigned char __c); @@ -126,7 +124,7 @@ vec_abs(vector signed int __a) { return __builtin_altivec_vmaxsw(__a, -__a); } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ static __inline__ vector signed long long __ATTRS_o_ai vec_abs(vector signed long long __a) { return __builtin_altivec_vmaxsd(__a, -__a); @@ -284,7 +282,7 @@ vec_add(vector unsigned int __a, vector bool int __b) { return __a + (vector unsigned int)__b; } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ static __inline__ vector signed long long __ATTRS_o_ai vec_add(vector signed long long __a, vector signed long long __b) { return __a + __b; @@ -295,6 +293,7 @@ vec_add(vector unsigned long long __a, vector unsigned long long __b) { return __a + __b; } +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_add(vector signed __int128 __a, vector signed __int128 __b) { return __a + __b; @@ -304,7 +303,37 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_add(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __a + __b; } -#endif // defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#endif + +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_add_u128(vector unsigned char __a, vector unsigned char __b) { + return __builtin_altivec_vadduqm(__a, __b); +} +#elif defined(__VSX__) +static __inline__ vector signed long long __ATTRS_o_ai +vec_add(vector signed long long __a, vector signed long long __b) { +#ifdef __LITTLE_ENDIAN__ + // Little endian systems on CPU's prior to Power8 don't really exist + // so scalarizing is fine. + return __a + __b; +#else + vector unsigned int __res = + (vector unsigned int)__a + (vector unsigned int)__b; + vector unsigned int __carry = __builtin_altivec_vaddcuw( + (vector unsigned int)__a, (vector unsigned int)__b); + __carry = __builtin_shufflevector((vector unsigned char)__carry, + (vector unsigned char)__carry, 0, 0, 0, 7, + 0, 0, 0, 0, 0, 0, 0, 15, 0, 0, 0, 0); + return (vector signed long long)(__res + __carry); +#endif +} + +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_add(vector unsigned long long __a, vector unsigned long long __b) { + return (vector unsigned long long)vec_add((vector signed long long)__a, + (vector signed long long)__b); +} +#endif // __POWER8_VECTOR__ static __inline__ vector float __ATTRS_o_ai vec_add(vector float __a, vector float __b) { @@ -320,7 +349,8 @@ static __inline__ vector double __ATTRS_o_ai vec_add(vector double __a, /* vec_adde */ -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_adde(vector signed __int128 __a, vector signed __int128 __b, vector signed __int128 __c) { @@ -334,6 +364,13 @@ vec_adde(vector unsigned __int128 __a, vector unsigned __int128 __b, } #endif +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_adde_u128(vector unsigned char __a, vector unsigned char __b, + vector unsigned char __c) { + return (vector unsigned char)__builtin_altivec_vaddeuqm(__a, __b, __c); +} +#endif + static __inline__ vector signed int __ATTRS_o_ai vec_adde(vector signed int __a, vector signed int __b, vector signed int __c) { @@ -352,7 +389,8 @@ vec_adde(vector unsigned int __a, vector unsigned int __b, /* vec_addec */ -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_addec(vector signed __int128 __a, vector signed __int128 __b, vector signed __int128 __c) { @@ -364,7 +402,15 @@ vec_addec(vector unsigned __int128 __a, vector unsigned __int128 __b, vector unsigned __int128 __c) { return __builtin_altivec_vaddecuq(__a, __b, __c); } +#endif +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_addec_u128(vector unsigned char __a, vector unsigned char __b, + vector unsigned char __c) { + return (vector unsigned char)__builtin_altivec_vaddecuq(__a, __b, __c); +} + +#ifdef __powerpc64__ static __inline__ vector signed int __ATTRS_o_ai vec_addec(vector signed int __a, vector signed int __b, vector signed int __c) { @@ -407,8 +453,8 @@ vec_addec(vector unsigned int __a, vector unsigned int __b, vector unsigned int ret = { __result[0], __result[1], __result[2], __result[3] }; return ret; } - -#endif +#endif // __powerpc64__ +#endif // __POWER8_VECTOR__ /* vec_vaddubm */ @@ -534,7 +580,8 @@ vec_addc(vector unsigned int __a, vector unsigned int __b) { return __builtin_altivec_vaddcuw(__a, __b); } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_addc(vector signed __int128 __a, vector signed __int128 __b) { return (vector signed __int128)__builtin_altivec_vaddcuq( @@ -545,6 +592,12 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_addc(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __builtin_altivec_vaddcuq(__a, __b); } +#endif + +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_addc_u128(vector unsigned char __a, vector unsigned char __b) { + return (vector unsigned char)__builtin_altivec_vaddcuq(__a, __b); +} #endif // defined(__POWER8_VECTOR__) && defined(__powerpc64__) /* vec_vaddcuw */ @@ -748,7 +801,8 @@ vec_vadduws(vector unsigned int __a, vector bool int __b) { return __builtin_altivec_vadduws(__a, (vector unsigned int)__b); } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) /* vec_vadduqm */ static __inline__ vector signed __int128 __ATTRS_o_ai @@ -1598,6 +1652,17 @@ static __inline__ vector double __ATTRS_o_ai vec_ceil(vector double __a) { } #endif +/* vec_roundp */ +static __inline__ vector float __ATTRS_o_ai vec_roundp(vector float __a) { + return vec_ceil(__a); +} + +#ifdef __VSX__ +static __inline__ vector double __ATTRS_o_ai vec_roundp(vector double __a) { + return vec_ceil(__a); +} +#endif + /* vec_vrfip */ static __inline__ vector float __attribute__((__always_inline__)) @@ -1690,7 +1755,31 @@ vec_cmpeq(vector bool long long __a, vector bool long long __b) { return (vector bool long long)__builtin_altivec_vcmpequd( (vector long long)__a, (vector long long)__b); } +#elif defined(__VSX__) +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpeq(vector signed long long __a, vector signed long long __b) { + vector bool int __wordcmp = + vec_cmpeq((vector signed int)__a, (vector signed int)__b); +#ifdef __LITTLE_ENDIAN__ + __wordcmp &= __builtin_shufflevector(__wordcmp, __wordcmp, 3, 0, 1, 2); + return (vector bool long long)__builtin_shufflevector(__wordcmp, __wordcmp, 1, + 1, 3, 3); +#else + __wordcmp &= __builtin_shufflevector(__wordcmp, __wordcmp, 1, 2, 3, 0); + return (vector bool long long)__builtin_shufflevector(__wordcmp, __wordcmp, 0, + 0, 2, 2); +#endif +} +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpeq(vector unsigned long long __a, vector unsigned long long __b) { + return vec_cmpeq((vector signed long long)__a, (vector signed long long)__b); +} + +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpeq(vector bool long long __a, vector bool long long __b) { + return vec_cmpeq((vector signed long long)__a, (vector signed long long)__b); +} #endif static __inline__ vector bool int __ATTRS_o_ai vec_cmpeq(vector float __a, @@ -1709,7 +1798,7 @@ vec_cmpeq(vector double __a, vector double __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector bool __int128 __ATTRS_o_ai vec_cmpeq(vector signed __int128 __a, vector signed __int128 __b) { return (vector bool __int128)__builtin_altivec_vcmpequq( @@ -1786,7 +1875,7 @@ vec_cmpne(vector float __a, vector float __b) { (vector int)__b); } -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector bool __int128 __ATTRS_o_ai vec_cmpne(vector unsigned __int128 __a, vector unsigned __int128 __b) { return (vector bool __int128) ~(__builtin_altivec_vcmpequq( @@ -1884,6 +1973,7 @@ vec_parity_lsbb(vector signed int __a) { return __builtin_altivec_vprtybw(__a); } +#ifdef __SIZEOF_INT128__ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_parity_lsbb(vector unsigned __int128 __a) { return __builtin_altivec_vprtybq(__a); @@ -1893,6 +1983,7 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_parity_lsbb(vector signed __int128 __a) { return __builtin_altivec_vprtybq(__a); } +#endif static __inline__ vector unsigned long long __ATTRS_o_ai vec_parity_lsbb(vector unsigned long long __a) { @@ -1976,6 +2067,24 @@ vec_cmpne(vector unsigned long long __a, vector unsigned long long __b) { return (vector bool long long) ~(__builtin_altivec_vcmpequd((vector long long)__a, (vector long long)__b)); } +#elif defined(__VSX__) +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpne(vector bool long long __a, vector bool long long __b) { + return (vector bool long long)~( + vec_cmpeq((vector signed long long)__a, (vector signed long long)__b)); +} + +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpne(vector signed long long __a, vector signed long long __b) { + return (vector bool long long)~( + vec_cmpeq((vector signed long long)__a, (vector signed long long)__b)); +} + +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpne(vector unsigned long long __a, vector unsigned long long __b) { + return (vector bool long long)~( + vec_cmpeq((vector signed long long)__a, (vector signed long long)__b)); +} #endif #ifdef __VSX__ @@ -2028,6 +2137,46 @@ static __inline__ vector bool long long __ATTRS_o_ai vec_cmpgt(vector unsigned long long __a, vector unsigned long long __b) { return (vector bool long long)__builtin_altivec_vcmpgtud(__a, __b); } +#elif defined(__VSX__) +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpgt(vector signed long long __a, vector signed long long __b) { + vector signed int __sgtw = (vector signed int)vec_cmpgt( + (vector signed int)__a, (vector signed int)__b); + vector unsigned int __ugtw = (vector unsigned int)vec_cmpgt( + (vector unsigned int)__a, (vector unsigned int)__b); + vector unsigned int __eqw = (vector unsigned int)vec_cmpeq( + (vector signed int)__a, (vector signed int)__b); +#ifdef __LITTLE_ENDIAN__ + __ugtw = __builtin_shufflevector(__ugtw, __ugtw, 3, 0, 1, 2) & __eqw; + __sgtw |= (vector signed int)__ugtw; + return (vector bool long long)__builtin_shufflevector(__sgtw, __sgtw, 1, 1, 3, + 3); +#else + __ugtw = __builtin_shufflevector(__ugtw, __ugtw, 1, 2, 3, 0) & __eqw; + __sgtw |= (vector signed int)__ugtw; + return (vector bool long long)__builtin_shufflevector(__sgtw, __sgtw, 0, 0, 2, + 2); +#endif +} + +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpgt(vector unsigned long long __a, vector unsigned long long __b) { + vector unsigned int __ugtw = (vector unsigned int)vec_cmpgt( + (vector unsigned int)__a, (vector unsigned int)__b); + vector unsigned int __eqw = (vector unsigned int)vec_cmpeq( + (vector signed int)__a, (vector signed int)__b); +#ifdef __LITTLE_ENDIAN__ + __eqw = __builtin_shufflevector(__ugtw, __ugtw, 3, 0, 1, 2) & __eqw; + __ugtw |= __eqw; + return (vector bool long long)__builtin_shufflevector(__ugtw, __ugtw, 1, 1, 3, + 3); +#else + __eqw = __builtin_shufflevector(__ugtw, __ugtw, 1, 2, 3, 0) & __eqw; + __ugtw |= __eqw; + return (vector bool long long)__builtin_shufflevector(__ugtw, __ugtw, 0, 0, 2, + 2); +#endif +} #endif static __inline__ vector bool int __ATTRS_o_ai vec_cmpgt(vector float __a, @@ -2046,7 +2195,7 @@ vec_cmpgt(vector double __a, vector double __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector bool __int128 __ATTRS_o_ai vec_cmpgt(vector signed __int128 __a, vector signed __int128 __b) { return (vector bool __int128)__builtin_altivec_vcmpgtsq( @@ -2106,9 +2255,7 @@ static __inline__ vector bool long long __ATTRS_o_ai vec_cmpge(vector double __a, vector double __b) { return (vector bool long long)__builtin_vsx_xvcmpgedp(__a, __b); } -#endif -#ifdef __POWER8_VECTOR__ static __inline__ vector bool long long __ATTRS_o_ai vec_cmpge(vector signed long long __a, vector signed long long __b) { return ~(vec_cmpgt(__b, __a)); @@ -2120,7 +2267,7 @@ vec_cmpge(vector unsigned long long __a, vector unsigned long long __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector bool __int128 __ATTRS_o_ai vec_cmpge(vector signed __int128 __a, vector signed __int128 __b) { return ~(vec_cmpgt(__b, __a)); @@ -2230,9 +2377,7 @@ static __inline__ vector bool long long __ATTRS_o_ai vec_cmple(vector double __a, vector double __b) { return vec_cmpge(__b, __a); } -#endif -#ifdef __POWER8_VECTOR__ static __inline__ vector bool long long __ATTRS_o_ai vec_cmple(vector signed long long __a, vector signed long long __b) { return vec_cmpge(__b, __a); @@ -2244,7 +2389,7 @@ vec_cmple(vector unsigned long long __a, vector unsigned long long __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector bool __int128 __ATTRS_o_ai vec_cmple(vector signed __int128 __a, vector signed __int128 __b) { return vec_cmpge(__b, __a); @@ -2300,7 +2445,7 @@ vec_cmplt(vector double __a, vector double __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector bool __int128 __ATTRS_o_ai vec_cmplt(vector signed __int128 __a, vector signed __int128 __b) { return vec_cmpgt(__b, __a); @@ -2312,7 +2457,7 @@ vec_cmplt(vector unsigned __int128 __a, vector unsigned __int128 __b) { } #endif -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ vector bool long long __ATTRS_o_ai vec_cmplt(vector signed long long __a, vector signed long long __b) { return vec_cmpgt(__b, __a); @@ -2322,7 +2467,9 @@ static __inline__ vector bool long long __ATTRS_o_ai vec_cmplt(vector unsigned long long __a, vector unsigned long long __b) { return vec_cmpgt(__b, __a); } +#endif +#ifdef __POWER8_VECTOR__ /* vec_popcnt */ static __inline__ vector signed char __ATTRS_o_ai @@ -2358,6 +2505,7 @@ vec_popcnt(vector unsigned long long __a) { return __builtin_altivec_vpopcntd(__a); } +#define vec_vclz vec_cntlz /* vec_cntlz */ static __inline__ vector signed char __ATTRS_o_ai @@ -2870,6 +3018,7 @@ static __inline__ vector float __ATTRS_o_ai vec_xl_len(const float *__a, size_t return (vector float)__builtin_vsx_lxvl(__a, (__b << 56)); } +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_xl_len(const signed __int128 *__a, size_t __b) { return (vector signed __int128)__builtin_vsx_lxvl(__a, (__b << 56)); @@ -2879,6 +3028,7 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_xl_len(const unsigned __int128 *__a, size_t __b) { return (vector unsigned __int128)__builtin_vsx_lxvl(__a, (__b << 56)); } +#endif static __inline__ vector signed long long __ATTRS_o_ai vec_xl_len(const signed long long *__a, size_t __b) { @@ -2946,6 +3096,7 @@ static __inline__ void __ATTRS_o_ai vec_xst_len(vector float __a, float *__b, return __builtin_vsx_stxvl((vector int)__a, __b, (__c << 56)); } +#ifdef __SIZEOF_INT128__ static __inline__ void __ATTRS_o_ai vec_xst_len(vector signed __int128 __a, signed __int128 *__b, size_t __c) { @@ -2957,6 +3108,7 @@ static __inline__ void __ATTRS_o_ai vec_xst_len(vector unsigned __int128 __a, size_t __c) { return __builtin_vsx_stxvl((vector int)__a, __b, (__c << 56)); } +#endif static __inline__ void __ATTRS_o_ai vec_xst_len(vector signed long long __a, signed long long *__b, @@ -2991,6 +3143,15 @@ static __inline__ void __ATTRS_o_ai vec_xst_len_r(vector unsigned char __a, #endif #endif +#if defined(__POWER9_VECTOR__) && defined(__powerpc64__) +#define __vec_ldrmb(PTR, CNT) vec_xl_len_r((const unsigned char *)(PTR), (CNT)) +#define __vec_strmb(PTR, CNT, VAL) \ + vec_xst_len_r((VAL), (unsigned char *)(PTR), (CNT)) +#else +#define __vec_ldrmb __builtin_vsx_ldrmb +#define __vec_strmb __builtin_vsx_strmb +#endif + /* vec_cpsgn */ #ifdef __VSX__ @@ -3008,6 +3169,23 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, /* vec_ctf */ #ifdef __VSX__ +// There are some functions that have different signatures with the XL compiler +// from those in Clang/GCC and documented in the PVIPR. This macro ensures that +// the XL-compatible signatures are used for those functions. +#ifdef __XL_COMPAT_ALTIVEC__ +#define vec_ctf(__a, __b) \ + _Generic((__a), vector int \ + : (vector float)__builtin_altivec_vcfsx((vector int)(__a), (__b)), \ + vector unsigned int \ + : (vector float)__builtin_altivec_vcfux((vector unsigned int)(__a), \ + (__b)), \ + vector unsigned long long \ + : (__builtin_vsx_xvcvuxdsp((vector unsigned long long)(__a)) * \ + (vector float)(vector unsigned)((0x7f - (__b)) << 23)), \ + vector signed long long \ + : (__builtin_vsx_xvcvsxdsp((vector signed long long)(__a)) * \ + (vector float)(vector unsigned)((0x7f - (__b)) << 23))) +#else // __XL_COMPAT_ALTIVEC__ #define vec_ctf(__a, __b) \ _Generic((__a), vector int \ : (vector float)__builtin_altivec_vcfsx((vector int)(__a), (__b)), \ @@ -3024,6 +3202,7 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, vector double) * \ (vector double)(vector unsigned long long)((0x3ffULL - (__b)) \ << 52))) +#endif // __XL_COMPAT_ALTIVEC__ #else #define vec_ctf(__a, __b) \ _Generic((__a), vector int \ @@ -3033,10 +3212,32 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, (__b))) #endif +/* vec_ctd */ +#ifdef __VSX__ +#define vec_ctd(__a, __b) \ + _Generic((__a), vector signed int \ + : (vec_doublee((vector signed int)(__a)) * \ + (vector double)(vector unsigned long long)((0x3ffULL - (__b)) \ + << 52)), \ + vector unsigned int \ + : (vec_doublee((vector unsigned int)(__a)) * \ + (vector double)(vector unsigned long long)((0x3ffULL - (__b)) \ + << 52)), \ + vector unsigned long long \ + : (__builtin_convertvector((vector unsigned long long)(__a), \ + vector double) * \ + (vector double)(vector unsigned long long)((0x3ffULL - (__b)) \ + << 52)), \ + vector signed long long \ + : (__builtin_convertvector((vector signed long long)(__a), \ + vector double) * \ + (vector double)(vector unsigned long long)((0x3ffULL - (__b)) \ + << 52))) +#endif // __VSX__ + /* vec_vcfsx */ #define vec_vcfux __builtin_altivec_vcfux - /* vec_vcfux */ #define vec_vcfsx(__a, __b) __builtin_altivec_vcfsx((vector int)(__a), (__b)) @@ -3044,6 +3245,19 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, /* vec_cts */ #ifdef __VSX__ +#ifdef __XL_COMPAT_ALTIVEC__ +#define vec_cts(__a, __b) \ + _Generic((__a), vector float \ + : __builtin_altivec_vctsxs((vector float)(__a), (__b)), \ + vector double \ + : __extension__({ \ + vector double __ret = \ + (vector double)(__a) * \ + (vector double)(vector unsigned long long)((0x3ffULL + (__b)) \ + << 52); \ + __builtin_vsx_xvcvdpsxws(__ret); \ + })) +#else // __XL_COMPAT_ALTIVEC__ #define vec_cts(__a, __b) \ _Generic((__a), vector float \ : __builtin_altivec_vctsxs((vector float)(__a), (__b)), \ @@ -3055,6 +3269,7 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, << 52); \ __builtin_convertvector(__ret, vector signed long long); \ })) +#endif // __XL_COMPAT_ALTIVEC__ #else #define vec_cts __builtin_altivec_vctsxs #endif @@ -3066,6 +3281,19 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, /* vec_ctu */ #ifdef __VSX__ +#ifdef __XL_COMPAT_ALTIVEC__ +#define vec_ctu(__a, __b) \ + _Generic((__a), vector float \ + : __builtin_altivec_vctuxs((vector float)(__a), (__b)), \ + vector double \ + : __extension__({ \ + vector double __ret = \ + (vector double)(__a) * \ + (vector double)(vector unsigned long long)((0x3ffULL + __b) \ + << 52); \ + __builtin_vsx_xvcvdpuxws(__ret); \ + })) +#else // __XL_COMPAT_ALTIVEC__ #define vec_ctu(__a, __b) \ _Generic((__a), vector float \ : __builtin_altivec_vctuxs((vector float)(__a), (__b)), \ @@ -3077,10 +3305,95 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, << 52); \ __builtin_convertvector(__ret, vector unsigned long long); \ })) +#endif // __XL_COMPAT_ALTIVEC__ #else #define vec_ctu __builtin_altivec_vctuxs #endif +#ifdef __LITTLE_ENDIAN__ +/* vec_ctsl */ + +#ifdef __VSX__ +#define vec_ctsl(__a, __b) \ + _Generic((__a), vector float \ + : __extension__({ \ + vector float __ret = \ + (vector float)(__a) * \ + (vector float)(vector unsigned)((0x7f + (__b)) << 23); \ + __builtin_vsx_xvcvspsxds( \ + __builtin_vsx_xxsldwi(__ret, __ret, 1)); \ + }), \ + vector double \ + : __extension__({ \ + vector double __ret = \ + (vector double)(__a) * \ + (vector double)(vector unsigned long long)((0x3ffULL + __b) \ + << 52); \ + __builtin_convertvector(__ret, vector signed long long); \ + })) + +/* vec_ctul */ + +#define vec_ctul(__a, __b) \ + _Generic((__a), vector float \ + : __extension__({ \ + vector float __ret = \ + (vector float)(__a) * \ + (vector float)(vector unsigned)((0x7f + (__b)) << 23); \ + __builtin_vsx_xvcvspuxds( \ + __builtin_vsx_xxsldwi(__ret, __ret, 1)); \ + }), \ + vector double \ + : __extension__({ \ + vector double __ret = \ + (vector double)(__a) * \ + (vector double)(vector unsigned long long)((0x3ffULL + __b) \ + << 52); \ + __builtin_convertvector(__ret, vector unsigned long long); \ + })) +#endif +#else // __LITTLE_ENDIAN__ +/* vec_ctsl */ + +#ifdef __VSX__ +#define vec_ctsl(__a, __b) \ + _Generic((__a), vector float \ + : __extension__({ \ + vector float __ret = \ + (vector float)(__a) * \ + (vector float)(vector unsigned)((0x7f + (__b)) << 23); \ + __builtin_vsx_xvcvspsxds(__ret); \ + }), \ + vector double \ + : __extension__({ \ + vector double __ret = \ + (vector double)(__a) * \ + (vector double)(vector unsigned long long)((0x3ffULL + __b) \ + << 52); \ + __builtin_convertvector(__ret, vector signed long long); \ + })) + +/* vec_ctul */ + +#define vec_ctul(__a, __b) \ + _Generic((__a), vector float \ + : __extension__({ \ + vector float __ret = \ + (vector float)(__a) * \ + (vector float)(vector unsigned)((0x7f + (__b)) << 23); \ + __builtin_vsx_xvcvspuxds(__ret); \ + }), \ + vector double \ + : __extension__({ \ + vector double __ret = \ + (vector double)(__a) * \ + (vector double)(vector unsigned long long)((0x3ffULL + __b) \ + << 52); \ + __builtin_convertvector(__ret, vector unsigned long long); \ + })) +#endif +#endif // __LITTLE_ENDIAN__ + /* vec_vctuxs */ #define vec_vctuxs __builtin_altivec_vctuxs @@ -3114,7 +3427,7 @@ vec_signextll(vector signed int __a) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_signextq(vector signed long long __a) { return __builtin_altivec_vextsd2q(__a); @@ -3399,6 +3712,15 @@ vec_doubleo(vector float __a) { return __builtin_vsx_xvcvspdp(vec_sld(__a, __a, 4)); #endif } + +/* vec_cvf */ +static __inline__ vector double __ATTRS_o_ai vec_cvf(vector float __a) { + return vec_doublee(__a); +} + +static __inline__ vector float __ATTRS_o_ai vec_cvf(vector double __a) { + return vec_floate(__a); +} #endif /* vec_div */ @@ -3481,6 +3803,7 @@ vec_dive(vector unsigned long long __a, vector unsigned long long __b) { return __builtin_altivec_vdiveud(__a, __b); } +#ifdef __SIZEOF_INT128__ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_dive(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __builtin_altivec_vdiveuq(__a, __b); @@ -3491,8 +3814,9 @@ vec_dive(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vdivesq(__a, __b); } #endif +#endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_div(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __a / __b; @@ -3661,6 +3985,17 @@ static __inline__ vector double __ATTRS_o_ai vec_floor(vector double __a) { } #endif +/* vec_roundm */ +static __inline__ vector float __ATTRS_o_ai vec_roundm(vector float __a) { + return vec_floor(__a); +} + +#ifdef __VSX__ +static __inline__ vector double __ATTRS_o_ai vec_roundm(vector double __a) { + return vec_floor(__a); +} +#endif + /* vec_vrfim */ static __inline__ vector float __attribute__((__always_inline__)) @@ -3671,251 +4006,251 @@ vec_vrfim(vector float __a) { /* vec_ld */ static __inline__ vector signed char __ATTRS_o_ai -vec_ld(int __a, const vector signed char *__b) { +vec_ld(long __a, const vector signed char *__b) { return (vector signed char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector signed char __ATTRS_o_ai -vec_ld(int __a, const signed char *__b) { +vec_ld(long __a, const signed char *__b) { return (vector signed char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_ld(int __a, const vector unsigned char *__b) { +vec_ld(long __a, const vector unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_ld(int __a, const unsigned char *__b) { +vec_ld(long __a, const unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector bool char __ATTRS_o_ai -vec_ld(int __a, const vector bool char *__b) { +vec_ld(long __a, const vector bool char *__b) { return (vector bool char)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_ld(int __a, +static __inline__ vector short __ATTRS_o_ai vec_ld(long __a, const vector short *__b) { return (vector short)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_ld(int __a, const short *__b) { +static __inline__ vector short __ATTRS_o_ai vec_ld(long __a, const short *__b) { return (vector short)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_ld(int __a, const vector unsigned short *__b) { +vec_ld(long __a, const vector unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_ld(int __a, const unsigned short *__b) { +vec_ld(long __a, const unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvx(__a, __b); } static __inline__ vector bool short __ATTRS_o_ai -vec_ld(int __a, const vector bool short *__b) { +vec_ld(long __a, const vector bool short *__b) { return (vector bool short)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector pixel __ATTRS_o_ai vec_ld(int __a, +static __inline__ vector pixel __ATTRS_o_ai vec_ld(long __a, const vector pixel *__b) { return (vector pixel)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_ld(int __a, +static __inline__ vector int __ATTRS_o_ai vec_ld(long __a, const vector int *__b) { return (vector int)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_ld(int __a, const int *__b) { +static __inline__ vector int __ATTRS_o_ai vec_ld(long __a, const int *__b) { return (vector int)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_ld(int __a, const vector unsigned int *__b) { +vec_ld(long __a, const vector unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_ld(int __a, const unsigned int *__b) { +vec_ld(long __a, const unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvx(__a, __b); } static __inline__ vector bool int __ATTRS_o_ai -vec_ld(int __a, const vector bool int *__b) { +vec_ld(long __a, const vector bool int *__b) { return (vector bool int)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_ld(int __a, +static __inline__ vector float __ATTRS_o_ai vec_ld(long __a, const vector float *__b) { return (vector float)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_ld(int __a, const float *__b) { +static __inline__ vector float __ATTRS_o_ai vec_ld(long __a, const float *__b) { return (vector float)__builtin_altivec_lvx(__a, __b); } /* vec_lvx */ static __inline__ vector signed char __ATTRS_o_ai -vec_lvx(int __a, const vector signed char *__b) { +vec_lvx(long __a, const vector signed char *__b) { return (vector signed char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector signed char __ATTRS_o_ai -vec_lvx(int __a, const signed char *__b) { +vec_lvx(long __a, const signed char *__b) { return (vector signed char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_lvx(int __a, const vector unsigned char *__b) { +vec_lvx(long __a, const vector unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_lvx(int __a, const unsigned char *__b) { +vec_lvx(long __a, const unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector bool char __ATTRS_o_ai -vec_lvx(int __a, const vector bool char *__b) { +vec_lvx(long __a, const vector bool char *__b) { return (vector bool char)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_lvx(int __a, +static __inline__ vector short __ATTRS_o_ai vec_lvx(long __a, const vector short *__b) { return (vector short)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_lvx(int __a, const short *__b) { +static __inline__ vector short __ATTRS_o_ai vec_lvx(long __a, const short *__b) { return (vector short)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_lvx(int __a, const vector unsigned short *__b) { +vec_lvx(long __a, const vector unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_lvx(int __a, const unsigned short *__b) { +vec_lvx(long __a, const unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvx(__a, __b); } static __inline__ vector bool short __ATTRS_o_ai -vec_lvx(int __a, const vector bool short *__b) { +vec_lvx(long __a, const vector bool short *__b) { return (vector bool short)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector pixel __ATTRS_o_ai vec_lvx(int __a, +static __inline__ vector pixel __ATTRS_o_ai vec_lvx(long __a, const vector pixel *__b) { return (vector pixel)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_lvx(int __a, +static __inline__ vector int __ATTRS_o_ai vec_lvx(long __a, const vector int *__b) { return (vector int)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_lvx(int __a, const int *__b) { +static __inline__ vector int __ATTRS_o_ai vec_lvx(long __a, const int *__b) { return (vector int)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_lvx(int __a, const vector unsigned int *__b) { +vec_lvx(long __a, const vector unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_lvx(int __a, const unsigned int *__b) { +vec_lvx(long __a, const unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvx(__a, __b); } static __inline__ vector bool int __ATTRS_o_ai -vec_lvx(int __a, const vector bool int *__b) { +vec_lvx(long __a, const vector bool int *__b) { return (vector bool int)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_lvx(int __a, +static __inline__ vector float __ATTRS_o_ai vec_lvx(long __a, const vector float *__b) { return (vector float)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_lvx(int __a, const float *__b) { +static __inline__ vector float __ATTRS_o_ai vec_lvx(long __a, const float *__b) { return (vector float)__builtin_altivec_lvx(__a, __b); } /* vec_lde */ static __inline__ vector signed char __ATTRS_o_ai -vec_lde(int __a, const signed char *__b) { +vec_lde(long __a, const signed char *__b) { return (vector signed char)__builtin_altivec_lvebx(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_lde(int __a, const unsigned char *__b) { +vec_lde(long __a, const unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvebx(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_lde(int __a, const short *__b) { +static __inline__ vector short __ATTRS_o_ai vec_lde(long __a, const short *__b) { return (vector short)__builtin_altivec_lvehx(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_lde(int __a, const unsigned short *__b) { +vec_lde(long __a, const unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvehx(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_lde(int __a, const int *__b) { +static __inline__ vector int __ATTRS_o_ai vec_lde(long __a, const int *__b) { return (vector int)__builtin_altivec_lvewx(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_lde(int __a, const unsigned int *__b) { +vec_lde(long __a, const unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvewx(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_lde(int __a, const float *__b) { +static __inline__ vector float __ATTRS_o_ai vec_lde(long __a, const float *__b) { return (vector float)__builtin_altivec_lvewx(__a, __b); } /* vec_lvebx */ static __inline__ vector signed char __ATTRS_o_ai -vec_lvebx(int __a, const signed char *__b) { +vec_lvebx(long __a, const signed char *__b) { return (vector signed char)__builtin_altivec_lvebx(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_lvebx(int __a, const unsigned char *__b) { +vec_lvebx(long __a, const unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvebx(__a, __b); } /* vec_lvehx */ -static __inline__ vector short __ATTRS_o_ai vec_lvehx(int __a, +static __inline__ vector short __ATTRS_o_ai vec_lvehx(long __a, const short *__b) { return (vector short)__builtin_altivec_lvehx(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_lvehx(int __a, const unsigned short *__b) { +vec_lvehx(long __a, const unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvehx(__a, __b); } /* vec_lvewx */ -static __inline__ vector int __ATTRS_o_ai vec_lvewx(int __a, const int *__b) { +static __inline__ vector int __ATTRS_o_ai vec_lvewx(long __a, const int *__b) { return (vector int)__builtin_altivec_lvewx(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_lvewx(int __a, const unsigned int *__b) { +vec_lvewx(long __a, const unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvewx(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_lvewx(int __a, +static __inline__ vector float __ATTRS_o_ai vec_lvewx(long __a, const float *__b) { return (vector float)__builtin_altivec_lvewx(__a, __b); } @@ -3923,179 +4258,179 @@ static __inline__ vector float __ATTRS_o_ai vec_lvewx(int __a, /* vec_ldl */ static __inline__ vector signed char __ATTRS_o_ai -vec_ldl(int __a, const vector signed char *__b) { +vec_ldl(long __a, const vector signed char *__b) { return (vector signed char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector signed char __ATTRS_o_ai -vec_ldl(int __a, const signed char *__b) { +vec_ldl(long __a, const signed char *__b) { return (vector signed char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_ldl(int __a, const vector unsigned char *__b) { +vec_ldl(long __a, const vector unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_ldl(int __a, const unsigned char *__b) { +vec_ldl(long __a, const unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector bool char __ATTRS_o_ai -vec_ldl(int __a, const vector bool char *__b) { +vec_ldl(long __a, const vector bool char *__b) { return (vector bool char)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_ldl(int __a, +static __inline__ vector short __ATTRS_o_ai vec_ldl(long __a, const vector short *__b) { return (vector short)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_ldl(int __a, const short *__b) { +static __inline__ vector short __ATTRS_o_ai vec_ldl(long __a, const short *__b) { return (vector short)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_ldl(int __a, const vector unsigned short *__b) { +vec_ldl(long __a, const vector unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_ldl(int __a, const unsigned short *__b) { +vec_ldl(long __a, const unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector bool short __ATTRS_o_ai -vec_ldl(int __a, const vector bool short *__b) { +vec_ldl(long __a, const vector bool short *__b) { return (vector bool short)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector pixel __ATTRS_o_ai vec_ldl(int __a, +static __inline__ vector pixel __ATTRS_o_ai vec_ldl(long __a, const vector pixel *__b) { return (vector pixel short)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_ldl(int __a, +static __inline__ vector int __ATTRS_o_ai vec_ldl(long __a, const vector int *__b) { return (vector int)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_ldl(int __a, const int *__b) { +static __inline__ vector int __ATTRS_o_ai vec_ldl(long __a, const int *__b) { return (vector int)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_ldl(int __a, const vector unsigned int *__b) { +vec_ldl(long __a, const vector unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_ldl(int __a, const unsigned int *__b) { +vec_ldl(long __a, const unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector bool int __ATTRS_o_ai -vec_ldl(int __a, const vector bool int *__b) { +vec_ldl(long __a, const vector bool int *__b) { return (vector bool int)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_ldl(int __a, +static __inline__ vector float __ATTRS_o_ai vec_ldl(long __a, const vector float *__b) { return (vector float)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_ldl(int __a, const float *__b) { +static __inline__ vector float __ATTRS_o_ai vec_ldl(long __a, const float *__b) { return (vector float)__builtin_altivec_lvxl(__a, __b); } /* vec_lvxl */ static __inline__ vector signed char __ATTRS_o_ai -vec_lvxl(int __a, const vector signed char *__b) { +vec_lvxl(long __a, const vector signed char *__b) { return (vector signed char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector signed char __ATTRS_o_ai -vec_lvxl(int __a, const signed char *__b) { +vec_lvxl(long __a, const signed char *__b) { return (vector signed char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_lvxl(int __a, const vector unsigned char *__b) { +vec_lvxl(long __a, const vector unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_lvxl(int __a, const unsigned char *__b) { +vec_lvxl(long __a, const unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector bool char __ATTRS_o_ai -vec_lvxl(int __a, const vector bool char *__b) { +vec_lvxl(long __a, const vector bool char *__b) { return (vector bool char)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_lvxl(int __a, +static __inline__ vector short __ATTRS_o_ai vec_lvxl(long __a, const vector short *__b) { return (vector short)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_lvxl(int __a, +static __inline__ vector short __ATTRS_o_ai vec_lvxl(long __a, const short *__b) { return (vector short)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_lvxl(int __a, const vector unsigned short *__b) { +vec_lvxl(long __a, const vector unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_lvxl(int __a, const unsigned short *__b) { +vec_lvxl(long __a, const unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector bool short __ATTRS_o_ai -vec_lvxl(int __a, const vector bool short *__b) { +vec_lvxl(long __a, const vector bool short *__b) { return (vector bool short)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector pixel __ATTRS_o_ai vec_lvxl(int __a, +static __inline__ vector pixel __ATTRS_o_ai vec_lvxl(long __a, const vector pixel *__b) { return (vector pixel)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_lvxl(int __a, +static __inline__ vector int __ATTRS_o_ai vec_lvxl(long __a, const vector int *__b) { return (vector int)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_lvxl(int __a, const int *__b) { +static __inline__ vector int __ATTRS_o_ai vec_lvxl(long __a, const int *__b) { return (vector int)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_lvxl(int __a, const vector unsigned int *__b) { +vec_lvxl(long __a, const vector unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_lvxl(int __a, const unsigned int *__b) { +vec_lvxl(long __a, const unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector bool int __ATTRS_o_ai -vec_lvxl(int __a, const vector bool int *__b) { +vec_lvxl(long __a, const vector bool int *__b) { return (vector bool int)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_lvxl(int __a, +static __inline__ vector float __ATTRS_o_ai vec_lvxl(long __a, const vector float *__b) { return (vector float)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_lvxl(int __a, +static __inline__ vector float __ATTRS_o_ai vec_lvxl(long __a, const float *__b) { return (vector float)__builtin_altivec_lvxl(__a, __b); } @@ -5695,7 +6030,7 @@ vec_msum(vector unsigned short __a, vector unsigned short __b, /* vec_msumc */ -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_msumc(vector unsigned long long __a, vector unsigned long long __b, vector unsigned __int128 __c) { @@ -5929,7 +6264,7 @@ vec_mule(vector unsigned int __a, vector unsigned int __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_mule(vector signed long long __a, vector signed long long __b) { #ifdef __LITTLE_ENDIAN__ @@ -6075,7 +6410,7 @@ vec_mulo(vector unsigned int __a, vector unsigned int __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_mulo(vector signed long long __a, vector signed long long __b) { #ifdef __LITTLE_ENDIAN__ @@ -7927,7 +8262,7 @@ vec_rl(vector unsigned long long __a, vector unsigned long long __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_rl(vector signed __int128 __a, vector unsigned __int128 __b) { return (__b << __a)|(__b >> ((__CHAR_BIT__ * sizeof(vector signed __int128)) - __a)); @@ -7954,7 +8289,7 @@ vec_rlmi(vector unsigned long long __a, vector unsigned long long __b, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_rlmi(vector unsigned __int128 __a, vector unsigned __int128 __b, vector unsigned __int128 __c) { @@ -7985,7 +8320,7 @@ vec_rlnm(vector unsigned long long __a, vector unsigned long long __b, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_rlnm(vector unsigned __int128 __a, vector unsigned __int128 __b, vector unsigned __int128 __c) { @@ -8060,11 +8395,7 @@ vec_vrlw(vector unsigned int __a, vector unsigned int __b) { /* vec_round */ static __inline__ vector float __ATTRS_o_ai vec_round(vector float __a) { -#ifdef __VSX__ - return __builtin_vsx_xvrspi(__a); -#else return __builtin_altivec_vrfin(__a); -#endif } #ifdef __VSX__ @@ -8082,6 +8413,16 @@ static __inline__ vector double __ATTRS_o_ai vec_rint(vector double __a) { return __builtin_vsx_xvrdpic(__a); } +/* vec_roundc */ + +static __inline__ vector float __ATTRS_o_ai vec_roundc(vector float __a) { + return __builtin_vsx_xvrspic(__a); +} + +static __inline__ vector double __ATTRS_o_ai vec_roundc(vector double __a) { + return __builtin_vsx_xvrdpic(__a); +} + /* vec_nearbyint */ static __inline__ vector float __ATTRS_o_ai vec_nearbyint(vector float __a) { @@ -8128,6 +8469,16 @@ static __inline__ vector double __ATTRS_o_ai vec_rsqrte(vector double __a) { } #endif +static vector float __ATTRS_o_ai vec_rsqrt(vector float __a) { + return __builtin_ppc_rsqrtf(__a); +} + +#ifdef __VSX__ +static vector double __ATTRS_o_ai vec_rsqrt(vector double __a) { + return __builtin_ppc_rsqrtd(__a); +} +#endif + /* vec_vrsqrtefp */ static __inline__ __vector float __attribute__((__always_inline__)) @@ -8484,6 +8835,52 @@ vec_sl(vector unsigned long long __a, vector unsigned long long __b) { __CHAR_BIT__)); } +static __inline__ vector long long __ATTRS_o_ai +vec_sl(vector long long __a, vector unsigned long long __b) { + return (vector long long)vec_sl((vector unsigned long long)__a, __b); +} +#else +static __inline__ vector unsigned char __ATTRS_o_ai +vec_vspltb(vector unsigned char __a, unsigned char __b); +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_sl(vector unsigned long long __a, vector unsigned long long __b) { + __b %= (vector unsigned long long)(sizeof(unsigned long long) * __CHAR_BIT__); + + // Big endian element one (the right doubleword) can be left shifted as-is. + // The other element needs to be swapped into the right doubleword and + // shifted. Then the right doublewords of the two result vectors are merged. + vector signed long long __rightelt = + (vector signed long long)__builtin_altivec_vslo((vector signed int)__a, + (vector signed int)__b); +#ifdef __LITTLE_ENDIAN__ + __rightelt = (vector signed long long)__builtin_altivec_vsl( + (vector signed int)__rightelt, + (vector signed int)vec_vspltb((vector unsigned char)__b, 0)); +#else + __rightelt = (vector signed long long)__builtin_altivec_vsl( + (vector signed int)__rightelt, + (vector signed int)vec_vspltb((vector unsigned char)__b, 15)); +#endif + __a = __builtin_shufflevector(__a, __a, 1, 0); + __b = __builtin_shufflevector(__b, __b, 1, 0); + vector signed long long __leftelt = + (vector signed long long)__builtin_altivec_vslo((vector signed int)__a, + (vector signed int)__b); +#ifdef __LITTLE_ENDIAN__ + __leftelt = (vector signed long long)__builtin_altivec_vsl( + (vector signed int)__leftelt, + (vector signed int)vec_vspltb((vector unsigned char)__b, 0)); + return (vector unsigned long long)__builtin_shufflevector(__rightelt, + __leftelt, 0, 2); +#else + __leftelt = (vector signed long long)__builtin_altivec_vsl( + (vector signed int)__leftelt, + (vector signed int)vec_vspltb((vector unsigned char)__b, 15)); + return (vector unsigned long long)__builtin_shufflevector(__leftelt, + __rightelt, 1, 3); +#endif +} + static __inline__ vector long long __ATTRS_o_ai vec_sl(vector long long __a, vector unsigned long long __b) { return (vector long long)vec_sl((vector unsigned long long)__a, __b); @@ -8850,6 +9247,11 @@ static __inline__ vector unsigned int __ATTRS_o_ai vec_sldw( return vec_sld(__a, __b, ((__c << 2) & 0x0F)); } +static __inline__ vector float __ATTRS_o_ai vec_sldw( + vector float __a, vector float __b, unsigned const int __c) { + return vec_sld(__a, __b, ((__c << 2) & 0x0F)); +} + #ifdef __VSX__ static __inline__ vector signed long long __ATTRS_o_ai vec_sldw(vector signed long long __a, vector signed long long __b, @@ -8862,6 +9264,11 @@ vec_sldw(vector unsigned long long __a, vector unsigned long long __b, unsigned const int __c) { return vec_sld(__a, __b, ((__c << 2) & 0x0F)); } + +static __inline__ vector double __ATTRS_o_ai vec_sldw( + vector double __a, vector double __b, unsigned const int __c) { + return vec_sld(__a, __b, ((__c << 2) & 0x0F)); +} #endif #ifdef __POWER9_VECTOR__ @@ -9939,6 +10346,50 @@ vec_sr(vector unsigned long long __a, vector unsigned long long __b) { __CHAR_BIT__)); } +static __inline__ vector long long __ATTRS_o_ai +vec_sr(vector long long __a, vector unsigned long long __b) { + return (vector long long)vec_sr((vector unsigned long long)__a, __b); +} +#else +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_sr(vector unsigned long long __a, vector unsigned long long __b) { + __b %= (vector unsigned long long)(sizeof(unsigned long long) * __CHAR_BIT__); + + // Big endian element zero (the left doubleword) can be right shifted as-is. + // However the shift amount must be in the right doubleword. + // The other element needs to be swapped into the left doubleword and + // shifted. Then the left doublewords of the two result vectors are merged. + vector unsigned long long __swapshift = + __builtin_shufflevector(__b, __b, 1, 0); + vector unsigned long long __leftelt = + (vector unsigned long long)__builtin_altivec_vsro( + (vector signed int)__a, (vector signed int)__swapshift); +#ifdef __LITTLE_ENDIAN__ + __leftelt = (vector unsigned long long)__builtin_altivec_vsr( + (vector signed int)__leftelt, + (vector signed int)vec_vspltb((vector unsigned char)__swapshift, 0)); +#else + __leftelt = (vector unsigned long long)__builtin_altivec_vsr( + (vector signed int)__leftelt, + (vector signed int)vec_vspltb((vector unsigned char)__swapshift, 15)); +#endif + __a = __builtin_shufflevector(__a, __a, 1, 0); + vector unsigned long long __rightelt = + (vector unsigned long long)__builtin_altivec_vsro((vector signed int)__a, + (vector signed int)__b); +#ifdef __LITTLE_ENDIAN__ + __rightelt = (vector unsigned long long)__builtin_altivec_vsr( + (vector signed int)__rightelt, + (vector signed int)vec_vspltb((vector unsigned char)__b, 0)); + return __builtin_shufflevector(__rightelt, __leftelt, 1, 3); +#else + __rightelt = (vector unsigned long long)__builtin_altivec_vsr( + (vector signed int)__rightelt, + (vector signed int)vec_vspltb((vector unsigned char)__b, 15)); + return __builtin_shufflevector(__leftelt, __rightelt, 0, 2); +#endif +} + static __inline__ vector long long __ATTRS_o_ai vec_sr(vector long long __a, vector unsigned long long __b) { return (vector long long)vec_sr((vector unsigned long long)__a, __b); @@ -10029,6 +10480,18 @@ static __inline__ vector unsigned long long __ATTRS_o_ai vec_sra(vector unsigned long long __a, vector unsigned long long __b) { return (vector unsigned long long)((vector signed long long)__a >> __b); } +#else +static __inline__ vector signed long long __ATTRS_o_ai +vec_sra(vector signed long long __a, vector unsigned long long __b) { + __b %= (vector unsigned long long)(sizeof(unsigned long long) * __CHAR_BIT__); + return __a >> __b; +} + +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_sra(vector unsigned long long __a, vector unsigned long long __b) { + __b %= (vector unsigned long long)(sizeof(unsigned long long) * __CHAR_BIT__); + return (vector unsigned long long)((vector signed long long)__a >> __b); +} #endif /* vec_vsrab */ @@ -10635,420 +11098,420 @@ static __inline__ vector float __ATTRS_o_ai vec_vsro(vector float __a, /* vec_st */ -static __inline__ void __ATTRS_o_ai vec_st(vector signed char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector signed char __a, long __b, vector signed char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector signed char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector signed char __a, long __b, signed char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector unsigned char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector unsigned char __a, long __b, vector unsigned char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector unsigned char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector unsigned char __a, long __b, unsigned char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool char __a, long __b, signed char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool char __a, long __b, unsigned char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool char __a, long __b, vector bool char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector short __a, long __b, vector short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector short __a, long __b, short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector unsigned short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector unsigned short __a, long __b, vector unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector unsigned short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector unsigned short __a, long __b, unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool short __a, long __b, short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool short __a, long __b, unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool short __a, long __b, vector bool short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector pixel __a, long __b, short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector pixel __a, long __b, unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector pixel __a, long __b, vector pixel *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector int __a, long __b, vector int *__c) { __builtin_altivec_stvx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector int __a, int __b, int *__c) { +static __inline__ void __ATTRS_o_ai vec_st(vector int __a, long __b, int *__c) { __builtin_altivec_stvx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector unsigned int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector unsigned int __a, long __b, vector unsigned int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector unsigned int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector unsigned int __a, long __b, unsigned int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool int __a, long __b, int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool int __a, long __b, unsigned int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool int __a, long __b, vector bool int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector float __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector float __a, long __b, vector float *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector float __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector float __a, long __b, float *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } /* vec_stvx */ -static __inline__ void __ATTRS_o_ai vec_stvx(vector signed char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector signed char __a, long __b, vector signed char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector signed char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector signed char __a, long __b, signed char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned char __a, long __b, vector unsigned char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned char __a, long __b, unsigned char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool char __a, long __b, signed char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool char __a, long __b, unsigned char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool char __a, long __b, vector bool char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector short __a, long __b, vector short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector short __a, long __b, short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned short __a, long __b, vector unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned short __a, long __b, unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool short __a, long __b, short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool short __a, long __b, unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool short __a, long __b, vector bool short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector pixel __a, long __b, short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector pixel __a, long __b, unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector pixel __a, long __b, vector pixel *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector int __a, long __b, vector int *__c) { __builtin_altivec_stvx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector int __a, long __b, int *__c) { __builtin_altivec_stvx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned int __a, long __b, vector unsigned int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned int __a, long __b, unsigned int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool int __a, long __b, int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool int __a, long __b, unsigned int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool int __a, long __b, vector bool int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector float __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector float __a, long __b, vector float *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector float __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector float __a, long __b, float *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } /* vec_ste */ -static __inline__ void __ATTRS_o_ai vec_ste(vector signed char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector signed char __a, long __b, signed char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector unsigned char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector unsigned char __a, long __b, unsigned char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector bool char __a, long __b, signed char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector bool char __a, long __b, unsigned char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector short __a, long __b, short *__c) { __builtin_altivec_stvehx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector unsigned short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector unsigned short __a, long __b, unsigned short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector bool short __a, long __b, short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector bool short __a, long __b, unsigned short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector pixel __a, long __b, short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector pixel __a, long __b, unsigned short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector int __a, int __b, int *__c) { +static __inline__ void __ATTRS_o_ai vec_ste(vector int __a, long __b, int *__c) { __builtin_altivec_stvewx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector unsigned int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector unsigned int __a, long __b, unsigned int *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector bool int __a, long __b, int *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector bool int __a, long __b, unsigned int *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector float __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector float __a, long __b, float *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } /* vec_stvebx */ -static __inline__ void __ATTRS_o_ai vec_stvebx(vector signed char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvebx(vector signed char __a, long __b, signed char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } static __inline__ void __ATTRS_o_ai vec_stvebx(vector unsigned char __a, - int __b, unsigned char *__c) { + long __b, unsigned char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvebx(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvebx(vector bool char __a, long __b, signed char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvebx(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvebx(vector bool char __a, long __b, unsigned char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } /* vec_stvehx */ -static __inline__ void __ATTRS_o_ai vec_stvehx(vector short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvehx(vector short __a, long __b, short *__c) { __builtin_altivec_stvehx(__a, __b, __c); } static __inline__ void __ATTRS_o_ai vec_stvehx(vector unsigned short __a, - int __b, unsigned short *__c) { + long __b, unsigned short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvehx(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvehx(vector bool short __a, long __b, short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvehx(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvehx(vector bool short __a, long __b, unsigned short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvehx(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvehx(vector pixel __a, long __b, short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvehx(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvehx(vector pixel __a, long __b, unsigned short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } /* vec_stvewx */ -static __inline__ void __ATTRS_o_ai vec_stvewx(vector int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvewx(vector int __a, long __b, int *__c) { __builtin_altivec_stvewx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvewx(vector unsigned int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvewx(vector unsigned int __a, long __b, unsigned int *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvewx(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvewx(vector bool int __a, long __b, int *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvewx(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvewx(vector bool int __a, long __b, unsigned int *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvewx(vector float __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvewx(vector float __a, long __b, float *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } @@ -11409,7 +11872,8 @@ vec_sub(vector unsigned int __a, vector bool int __b) { return __a - (vector unsigned int)__b; } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_sub(vector signed __int128 __a, vector signed __int128 __b) { return __a - __b; @@ -11419,7 +11883,8 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_sub(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __a - __b; } -#endif // defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#endif // defined(__POWER8_VECTOR__) && defined(__powerpc64__) && + // defined(__SIZEOF_INT128__) #ifdef __VSX__ static __inline__ vector signed long long __ATTRS_o_ai @@ -11567,7 +12032,8 @@ vec_subc(vector unsigned int __a, vector unsigned int __b) { return __builtin_altivec_vsubcuw(__a, __b); } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ +#ifdef __SIZEOF_INT128__ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_subc(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __builtin_altivec_vsubcuq(__a, __b); @@ -11577,7 +12043,13 @@ static __inline__ vector signed __int128 __ATTRS_o_ai vec_subc(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vsubcuq(__a, __b); } -#endif // defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#endif + +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_subc_u128(vector unsigned char __a, vector unsigned char __b) { + return (vector unsigned char)__builtin_altivec_vsubcuq(__a, __b); +} +#endif // __POWER8_VECTOR__ /* vec_vsubcuw */ @@ -11780,9 +12252,10 @@ vec_vsubuws(vector unsigned int __a, vector bool int __b) { return __builtin_altivec_vsubuws(__a, (vector unsigned int)__b); } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ /* vec_vsubuqm */ +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_vsubuqm(vector signed __int128 __a, vector signed __int128 __b) { return __a - __b; @@ -11792,10 +12265,16 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_vsubuqm(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __a - __b; } +#endif + +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_sub_u128(vector unsigned char __a, vector unsigned char __b) { + return __builtin_altivec_vsubuqm(__a, __b); +} /* vec_vsubeuqm */ - +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_vsubeuqm(vector signed __int128 __a, vector signed __int128 __b, vector signed __int128 __c) { @@ -11819,9 +12298,17 @@ vec_sube(vector unsigned __int128 __a, vector unsigned __int128 __b, vector unsigned __int128 __c) { return __builtin_altivec_vsubeuqm(__a, __b, __c); } +#endif + +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_sube_u128(vector unsigned char __a, vector unsigned char __b, + vector unsigned char __c) { + return (vector unsigned char)__builtin_altivec_vsubeuqm(__a, __b, __c); +} /* vec_vsubcuq */ +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_vsubcuq(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vsubcuq(__a, __b); @@ -11845,7 +12332,9 @@ vec_vsubecuq(vector unsigned __int128 __a, vector unsigned __int128 __b, vector unsigned __int128 __c) { return __builtin_altivec_vsubecuq(__a, __b, __c); } +#endif +#ifdef __powerpc64__ static __inline__ vector signed int __ATTRS_o_ai vec_subec(vector signed int __a, vector signed int __b, vector signed int __c) { @@ -11857,7 +12346,9 @@ vec_subec(vector unsigned int __a, vector unsigned int __b, vector unsigned int __c) { return vec_addec(__a, ~__b, __c); } +#endif +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_subec(vector signed __int128 __a, vector signed __int128 __b, vector signed __int128 __c) { @@ -11869,7 +12360,14 @@ vec_subec(vector unsigned __int128 __a, vector unsigned __int128 __b, vector unsigned __int128 __c) { return __builtin_altivec_vsubecuq(__a, __b, __c); } -#endif // defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#endif + +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_subec_u128(vector unsigned char __a, vector unsigned char __b, + vector unsigned char __c) { + return (vector unsigned char)__builtin_altivec_vsubecuq(__a, __b, __c); +} +#endif // __POWER8_VECTOR__ static __inline__ vector signed int __ATTRS_o_ai vec_sube(vector signed int __a, vector signed int __b, @@ -12012,6 +12510,17 @@ static __inline__ vector double __ATTRS_o_ai vec_trunc(vector double __a) { } #endif +/* vec_roundz */ +static __inline__ vector float __ATTRS_o_ai vec_roundz(vector float __a) { + return vec_trunc(__a); +} + +#ifdef __VSX__ +static __inline__ vector double __ATTRS_o_ai vec_roundz(vector double __a) { + return vec_trunc(__a); +} +#endif + /* vec_vrfiz */ static __inline__ vector float __attribute__((__always_inline__)) @@ -12023,6 +12532,13 @@ vec_vrfiz(vector float __a) { /* The vector unpack instructions all have a big-endian bias, so for little endian we must reverse the meanings of "high" and "low." */ +#ifdef __LITTLE_ENDIAN__ +#define vec_vupkhpx(__a) __builtin_altivec_vupklpx((vector short)(__a)) +#define vec_vupklpx(__a) __builtin_altivec_vupkhpx((vector short)(__a)) +#else +#define vec_vupkhpx(__a) __builtin_altivec_vupkhpx((vector short)(__a)) +#define vec_vupklpx(__a) __builtin_altivec_vupklpx((vector short)(__a)) +#endif static __inline__ vector short __ATTRS_o_ai vec_unpackh(vector signed char __a) { @@ -12558,6 +13074,16 @@ static __inline__ void __ATTRS_o_ai vec_vsx_st(vector unsigned char __a, #ifdef __VSX__ #define vec_xxpermdi __builtin_vsx_xxpermdi #define vec_xxsldwi __builtin_vsx_xxsldwi +#define vec_permi(__a, __b, __c) \ + _Generic((__a), vector signed long long \ + : __builtin_shufflevector((__a), (__b), (((__c) >> 1) & 0x1), \ + (((__c)&0x1) + 2)), \ + vector unsigned long long \ + : __builtin_shufflevector((__a), (__b), (((__c) >> 1) & 0x1), \ + (((__c)&0x1) + 2)), \ + vector double \ + : __builtin_shufflevector((__a), (__b), (((__c) >> 1) & 0x1), \ + (((__c)&0x1) + 2))) #endif /* vec_xor */ @@ -12915,73 +13441,75 @@ vec_vxor(vector bool long long __a, vector bool long long __b) { /* vec_extract */ static __inline__ signed char __ATTRS_o_ai vec_extract(vector signed char __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0xf]; } static __inline__ unsigned char __ATTRS_o_ai -vec_extract(vector unsigned char __a, int __b) { - return __a[__b]; +vec_extract(vector unsigned char __a, unsigned int __b) { + return __a[__b & 0xf]; } static __inline__ unsigned char __ATTRS_o_ai vec_extract(vector bool char __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0xf]; } static __inline__ signed short __ATTRS_o_ai vec_extract(vector signed short __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0x7]; } static __inline__ unsigned short __ATTRS_o_ai -vec_extract(vector unsigned short __a, int __b) { - return __a[__b]; +vec_extract(vector unsigned short __a, unsigned int __b) { + return __a[__b & 0x7]; } static __inline__ unsigned short __ATTRS_o_ai vec_extract(vector bool short __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0x7]; } static __inline__ signed int __ATTRS_o_ai vec_extract(vector signed int __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0x3]; } static __inline__ unsigned int __ATTRS_o_ai vec_extract(vector unsigned int __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0x3]; } static __inline__ unsigned int __ATTRS_o_ai vec_extract(vector bool int __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0x3]; } #ifdef __VSX__ static __inline__ signed long long __ATTRS_o_ai -vec_extract(vector signed long long __a, int __b) { - return __a[__b]; +vec_extract(vector signed long long __a, unsigned int __b) { + return __a[__b & 0x1]; } static __inline__ unsigned long long __ATTRS_o_ai -vec_extract(vector unsigned long long __a, int __b) { - return __a[__b]; +vec_extract(vector unsigned long long __a, unsigned int __b) { + return __a[__b & 0x1]; } static __inline__ unsigned long long __ATTRS_o_ai -vec_extract(vector bool long long __a, int __b) { - return __a[__b]; +vec_extract(vector bool long long __a, unsigned int __b) { + return __a[__b & 0x1]; } -static __inline__ double __ATTRS_o_ai vec_extract(vector double __a, int __b) { - return __a[__b]; +static __inline__ double __ATTRS_o_ai vec_extract(vector double __a, + unsigned int __b) { + return __a[__b & 0x1]; } #endif -static __inline__ float __ATTRS_o_ai vec_extract(vector float __a, int __b) { - return __a[__b]; +static __inline__ float __ATTRS_o_ai vec_extract(vector float __a, + unsigned int __b) { + return __a[__b & 0x3]; } #ifdef __POWER9_VECTOR__ @@ -14022,49 +14550,71 @@ static __inline__ void __ATTRS_o_ai vec_stvrxl(vector float __a, int __b, static __inline__ vector signed char __ATTRS_o_ai vec_promote(signed char __a, int __b) { vector signed char __res = (vector signed char)(0); - __res[__b] = __a; + __res[__b & 0x7] = __a; return __res; } static __inline__ vector unsigned char __ATTRS_o_ai vec_promote(unsigned char __a, int __b) { vector unsigned char __res = (vector unsigned char)(0); - __res[__b] = __a; + __res[__b & 0x7] = __a; return __res; } static __inline__ vector short __ATTRS_o_ai vec_promote(short __a, int __b) { vector short __res = (vector short)(0); - __res[__b] = __a; + __res[__b & 0x7] = __a; return __res; } static __inline__ vector unsigned short __ATTRS_o_ai vec_promote(unsigned short __a, int __b) { vector unsigned short __res = (vector unsigned short)(0); - __res[__b] = __a; + __res[__b & 0x7] = __a; return __res; } static __inline__ vector int __ATTRS_o_ai vec_promote(int __a, int __b) { vector int __res = (vector int)(0); - __res[__b] = __a; + __res[__b & 0x3] = __a; return __res; } static __inline__ vector unsigned int __ATTRS_o_ai vec_promote(unsigned int __a, int __b) { vector unsigned int __res = (vector unsigned int)(0); - __res[__b] = __a; + __res[__b & 0x3] = __a; return __res; } static __inline__ vector float __ATTRS_o_ai vec_promote(float __a, int __b) { vector float __res = (vector float)(0); - __res[__b] = __a; + __res[__b & 0x3] = __a; return __res; } +#ifdef __VSX__ +static __inline__ vector double __ATTRS_o_ai vec_promote(double __a, int __b) { + vector double __res = (vector double)(0); + __res[__b & 0x1] = __a; + return __res; +} + +static __inline__ vector signed long long __ATTRS_o_ai +vec_promote(signed long long __a, int __b) { + vector signed long long __res = (vector signed long long)(0); + __res[__b & 0x1] = __a; + return __res; +} + +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_promote(unsigned long long __a, int __b) { + vector unsigned long long __res = (vector unsigned long long)(0); + __res[__b & 0x1] = __a; + return __res; +} +#endif + /* vec_splats */ static __inline__ vector signed char __ATTRS_o_ai vec_splats(signed char __a) { @@ -14105,7 +14655,8 @@ vec_splats(unsigned long long __a) { return (vector unsigned long long)(__a); } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_splats(signed __int128 __a) { return (vector signed __int128)(__a); @@ -14258,7 +14809,7 @@ static __inline__ int __ATTRS_o_ai vec_all_eq(vector bool int __a, (vector int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_all_eq(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpequd_p(__CR6_LT, __a, __b); @@ -14316,7 +14867,7 @@ static __inline__ int __ATTRS_o_ai vec_all_eq(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_all_eq(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpequq_p(__CR6_LT, __a, __b); @@ -14352,8 +14903,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ge(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_EQ, (vector unsigned char)__b, - (vector unsigned char)__a); + return __builtin_altivec_vcmpgtsb_p(__CR6_EQ, __b, (vector signed char)__a); } static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool char __a, @@ -14390,8 +14940,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ge(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_EQ, (vector unsigned short)__b, - (vector unsigned short)__a); + return __builtin_altivec_vcmpgtsh_p(__CR6_EQ, __b, (vector signed short)__a); } static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool short __a, @@ -14427,8 +14976,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ge(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_EQ, (vector unsigned int)__b, - (vector unsigned int)__a); + return __builtin_altivec_vcmpgtsw_p(__CR6_EQ, __b, (vector signed int)__a); } static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool int __a, @@ -14442,7 +14990,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool int __a, (vector unsigned int)__a); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_all_ge(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_EQ, __b, __a); @@ -14466,8 +15014,8 @@ static __inline__ int __ATTRS_o_ai vec_all_ge(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_EQ, (vector unsigned long long)__b, - (vector unsigned long long)__a); + return __builtin_altivec_vcmpgtsd_p(__CR6_EQ, __b, + (vector signed long long)__a); } static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool long long __a, @@ -14499,7 +15047,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ge(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_all_ge(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_EQ, __b, __a); @@ -14535,8 +15083,7 @@ static __inline__ int __ATTRS_o_ai vec_all_gt(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_LT, (vector unsigned char)__a, - (vector unsigned char)__b); + return __builtin_altivec_vcmpgtsb_p(__CR6_LT, (vector signed char)__a, __b); } static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool char __a, @@ -14573,8 +15120,7 @@ static __inline__ int __ATTRS_o_ai vec_all_gt(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_LT, (vector unsigned short)__a, - (vector unsigned short)__b); + return __builtin_altivec_vcmpgtsh_p(__CR6_LT, (vector signed short)__a, __b); } static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool short __a, @@ -14610,8 +15156,7 @@ static __inline__ int __ATTRS_o_ai vec_all_gt(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_LT, (vector unsigned int)__a, - (vector unsigned int)__b); + return __builtin_altivec_vcmpgtsw_p(__CR6_LT, (vector signed int)__a, __b); } static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool int __a, @@ -14625,7 +15170,7 @@ static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool int __a, (vector unsigned int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_all_gt(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_LT, __a, __b); @@ -14649,8 +15194,8 @@ static __inline__ int __ATTRS_o_ai vec_all_gt(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_LT, (vector unsigned long long)__a, - (vector unsigned long long)__b); + return __builtin_altivec_vcmpgtsd_p(__CR6_LT, (vector signed long long)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool long long __a, @@ -14682,7 +15227,7 @@ static __inline__ int __ATTRS_o_ai vec_all_gt(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_all_gt(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_LT, __a, __b); @@ -14725,8 +15270,7 @@ static __inline__ int __ATTRS_o_ai vec_all_le(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_all_le(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_EQ, (vector unsigned char)__a, - (vector unsigned char)__b); + return __builtin_altivec_vcmpgtsb_p(__CR6_EQ, (vector signed char)__a, __b); } static __inline__ int __ATTRS_o_ai vec_all_le(vector bool char __a, @@ -14763,8 +15307,7 @@ static __inline__ int __ATTRS_o_ai vec_all_le(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_all_le(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_EQ, (vector unsigned short)__a, - (vector unsigned short)__b); + return __builtin_altivec_vcmpgtsh_p(__CR6_EQ, (vector signed short)__a, __b); } static __inline__ int __ATTRS_o_ai vec_all_le(vector bool short __a, @@ -14800,8 +15343,7 @@ static __inline__ int __ATTRS_o_ai vec_all_le(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_all_le(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_EQ, (vector unsigned int)__a, - (vector unsigned int)__b); + return __builtin_altivec_vcmpgtsw_p(__CR6_EQ, (vector signed int)__a, __b); } static __inline__ int __ATTRS_o_ai vec_all_le(vector bool int __a, @@ -14815,7 +15357,7 @@ static __inline__ int __ATTRS_o_ai vec_all_le(vector bool int __a, (vector unsigned int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_all_le(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_EQ, __a, __b); @@ -14840,8 +15382,8 @@ static __inline__ int __ATTRS_o_ai vec_all_le(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_all_le(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_EQ, (vector unsigned long long)__a, - (vector unsigned long long)__b); + return __builtin_altivec_vcmpgtsd_p(__CR6_EQ, (vector signed long long)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_all_le(vector bool long long __a, @@ -14873,7 +15415,7 @@ static __inline__ int __ATTRS_o_ai vec_all_le(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_all_le(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_EQ, __a, __b); @@ -14909,8 +15451,7 @@ static __inline__ int __ATTRS_o_ai vec_all_lt(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_LT, (vector unsigned char)__b, - (vector unsigned char)__a); + return __builtin_altivec_vcmpgtsb_p(__CR6_LT, __b, (vector signed char)__a); } static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool char __a, @@ -14947,8 +15488,7 @@ static __inline__ int __ATTRS_o_ai vec_all_lt(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_LT, (vector unsigned short)__b, - (vector unsigned short)__a); + return __builtin_altivec_vcmpgtsh_p(__CR6_LT, __b, (vector signed short)__a); } static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool short __a, @@ -14984,8 +15524,7 @@ static __inline__ int __ATTRS_o_ai vec_all_lt(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_LT, (vector unsigned int)__b, - (vector unsigned int)__a); + return __builtin_altivec_vcmpgtsw_p(__CR6_LT, __b, (vector signed int)__a); } static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool int __a, @@ -14999,7 +15538,7 @@ static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool int __a, (vector unsigned int)__a); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_all_lt(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_LT, __b, __a); @@ -15024,8 +15563,8 @@ static __inline__ int __ATTRS_o_ai vec_all_lt(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_LT, (vector unsigned long long)__b, - (vector unsigned long long)__a); + return __builtin_altivec_vcmpgtsd_p(__CR6_LT, __b, + (vector signed long long)__a); } static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool long long __a, @@ -15057,7 +15596,7 @@ static __inline__ int __ATTRS_o_ai vec_all_lt(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_all_lt(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_LT, __b, __a); @@ -15214,7 +15753,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ne(vector bool int __a, (vector int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_all_ne(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpequd_p(__CR6_EQ, __a, __b); @@ -15273,7 +15812,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ne(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_all_ne(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpequq_p(__CR6_EQ, __a, __b); @@ -15323,25 +15862,57 @@ static __inline__ int __ATTRS_o_ai vec_all_ngt(vector double __a, /* vec_all_nle */ -static __inline__ int __attribute__((__always_inline__)) +static __inline__ int __ATTRS_o_ai vec_all_nle(vector float __a, vector float __b) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpgesp_p(__CR6_EQ, __b, __a); +#else return __builtin_altivec_vcmpgefp_p(__CR6_EQ, __b, __a); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_all_nle(vector double __a, + vector double __b) { + return __builtin_vsx_xvcmpgedp_p(__CR6_EQ, __b, __a); +} +#endif + /* vec_all_nlt */ -static __inline__ int __attribute__((__always_inline__)) +static __inline__ int __ATTRS_o_ai vec_all_nlt(vector float __a, vector float __b) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpgtsp_p(__CR6_EQ, __b, __a); +#else return __builtin_altivec_vcmpgtfp_p(__CR6_EQ, __b, __a); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_all_nlt(vector double __a, + vector double __b) { + return __builtin_vsx_xvcmpgtdp_p(__CR6_EQ, __b, __a); +} +#endif + /* vec_all_numeric */ -static __inline__ int __attribute__((__always_inline__)) +static __inline__ int __ATTRS_o_ai vec_all_numeric(vector float __a) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpeqsp_p(__CR6_LT, __a, __a); +#else return __builtin_altivec_vcmpeqfp_p(__CR6_LT, __a, __a); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_all_numeric(vector double __a) { + return __builtin_vsx_xvcmpeqdp_p(__CR6_LT, __a, __a); +} +#endif + /* vec_any_eq */ static __inline__ int __ATTRS_o_ai vec_any_eq(vector signed char __a, @@ -15471,7 +16042,7 @@ static __inline__ int __ATTRS_o_ai vec_any_eq(vector bool int __a, (vector int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_any_eq(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpequd_p(__CR6_EQ_REV, __a, __b); @@ -15530,7 +16101,7 @@ static __inline__ int __ATTRS_o_ai vec_any_eq(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_any_eq(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpequq_p(__CR6_EQ_REV, __a, __b); @@ -15568,8 +16139,8 @@ static __inline__ int __ATTRS_o_ai vec_any_ge(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_LT_REV, (vector unsigned char)__b, - (vector unsigned char)__a); + return __builtin_altivec_vcmpgtsb_p(__CR6_LT_REV, __b, + (vector signed char)__a); } static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool char __a, @@ -15607,8 +16178,8 @@ static __inline__ int __ATTRS_o_ai vec_any_ge(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_LT_REV, (vector unsigned short)__b, - (vector unsigned short)__a); + return __builtin_altivec_vcmpgtsh_p(__CR6_LT_REV, __b, + (vector signed short)__a); } static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool short __a, @@ -15645,8 +16216,8 @@ static __inline__ int __ATTRS_o_ai vec_any_ge(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_LT_REV, (vector unsigned int)__b, - (vector unsigned int)__a); + return __builtin_altivec_vcmpgtsw_p(__CR6_LT_REV, __b, + (vector signed int)__a); } static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool int __a, @@ -15661,7 +16232,7 @@ static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool int __a, (vector unsigned int)__a); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_any_ge(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_LT_REV, __b, __a); @@ -15686,9 +16257,8 @@ static __inline__ int __ATTRS_o_ai vec_any_ge(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_LT_REV, - (vector unsigned long long)__b, - (vector unsigned long long)__a); + return __builtin_altivec_vcmpgtsd_p(__CR6_LT_REV, __b, + (vector signed long long)__a); } static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool long long __a, @@ -15721,7 +16291,7 @@ static __inline__ int __ATTRS_o_ai vec_any_ge(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_any_ge(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_LT_REV, __b, __a); @@ -15759,8 +16329,8 @@ static __inline__ int __ATTRS_o_ai vec_any_gt(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_EQ_REV, (vector unsigned char)__a, - (vector unsigned char)__b); + return __builtin_altivec_vcmpgtsb_p(__CR6_EQ_REV, (vector signed char)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool char __a, @@ -15798,8 +16368,8 @@ static __inline__ int __ATTRS_o_ai vec_any_gt(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_EQ_REV, (vector unsigned short)__a, - (vector unsigned short)__b); + return __builtin_altivec_vcmpgtsh_p(__CR6_EQ_REV, (vector signed short)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool short __a, @@ -15836,8 +16406,8 @@ static __inline__ int __ATTRS_o_ai vec_any_gt(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_EQ_REV, (vector unsigned int)__a, - (vector unsigned int)__b); + return __builtin_altivec_vcmpgtsw_p(__CR6_EQ_REV, (vector signed int)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool int __a, @@ -15852,7 +16422,7 @@ static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool int __a, (vector unsigned int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_any_gt(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_EQ_REV, __a, __b); @@ -15877,9 +16447,8 @@ static __inline__ int __ATTRS_o_ai vec_any_gt(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_EQ_REV, - (vector unsigned long long)__a, - (vector unsigned long long)__b); + return __builtin_altivec_vcmpgtsd_p(__CR6_EQ_REV, + (vector signed long long)__a, __b); } static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool long long __a, @@ -15912,7 +16481,7 @@ static __inline__ int __ATTRS_o_ai vec_any_gt(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_any_gt(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_EQ_REV, __a, __b); @@ -15950,8 +16519,8 @@ static __inline__ int __ATTRS_o_ai vec_any_le(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_any_le(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_LT_REV, (vector unsigned char)__a, - (vector unsigned char)__b); + return __builtin_altivec_vcmpgtsb_p(__CR6_LT_REV, (vector signed char)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_any_le(vector bool char __a, @@ -15989,8 +16558,8 @@ static __inline__ int __ATTRS_o_ai vec_any_le(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_any_le(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_LT_REV, (vector unsigned short)__a, - (vector unsigned short)__b); + return __builtin_altivec_vcmpgtsh_p(__CR6_LT_REV, (vector signed short)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_any_le(vector bool short __a, @@ -16027,8 +16596,8 @@ static __inline__ int __ATTRS_o_ai vec_any_le(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_any_le(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_LT_REV, (vector unsigned int)__a, - (vector unsigned int)__b); + return __builtin_altivec_vcmpgtsw_p(__CR6_LT_REV, (vector signed int)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_any_le(vector bool int __a, @@ -16043,7 +16612,7 @@ static __inline__ int __ATTRS_o_ai vec_any_le(vector bool int __a, (vector unsigned int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_any_le(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_LT_REV, __a, __b); @@ -16068,9 +16637,8 @@ static __inline__ int __ATTRS_o_ai vec_any_le(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_any_le(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_LT_REV, - (vector unsigned long long)__a, - (vector unsigned long long)__b); + return __builtin_altivec_vcmpgtsd_p(__CR6_LT_REV, + (vector signed long long)__a, __b); } static __inline__ int __ATTRS_o_ai vec_any_le(vector bool long long __a, @@ -16103,7 +16671,7 @@ static __inline__ int __ATTRS_o_ai vec_any_le(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_any_le(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_LT_REV, __a, __b); @@ -16141,8 +16709,8 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_EQ_REV, (vector unsigned char)__b, - (vector unsigned char)__a); + return __builtin_altivec_vcmpgtsb_p(__CR6_EQ_REV, __b, + (vector signed char)__a); } static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool char __a, @@ -16180,8 +16748,8 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_EQ_REV, (vector unsigned short)__b, - (vector unsigned short)__a); + return __builtin_altivec_vcmpgtsh_p(__CR6_EQ_REV, __b, + (vector signed short)__a); } static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool short __a, @@ -16218,8 +16786,8 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_EQ_REV, (vector unsigned int)__b, - (vector unsigned int)__a); + return __builtin_altivec_vcmpgtsw_p(__CR6_EQ_REV, __b, + (vector signed int)__a); } static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool int __a, @@ -16234,7 +16802,7 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool int __a, (vector unsigned int)__a); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_any_lt(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_EQ_REV, __b, __a); @@ -16259,9 +16827,8 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_EQ_REV, - (vector unsigned long long)__b, - (vector unsigned long long)__a); + return __builtin_altivec_vcmpgtsd_p(__CR6_EQ_REV, __b, + (vector signed long long)__a); } static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool long long __a, @@ -16294,7 +16861,7 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_any_lt(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_EQ_REV, __b, __a); @@ -16308,10 +16875,18 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector unsigned __int128 __a, /* vec_any_nan */ -static __inline__ int __attribute__((__always_inline__)) -vec_any_nan(vector float __a) { +static __inline__ int __ATTRS_o_ai vec_any_nan(vector float __a) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpeqsp_p(__CR6_LT_REV, __a, __a); +#else return __builtin_altivec_vcmpeqfp_p(__CR6_LT_REV, __a, __a); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_any_nan(vector double __a) { + return __builtin_vsx_xvcmpeqdp_p(__CR6_LT_REV, __a, __a); +} +#endif /* vec_any_ne */ @@ -16442,7 +17017,7 @@ static __inline__ int __ATTRS_o_ai vec_any_ne(vector bool int __a, (vector int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_any_ne(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpequd_p(__CR6_LT_REV, __a, __b); @@ -16501,7 +17076,7 @@ static __inline__ int __ATTRS_o_ai vec_any_ne(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_any_ne(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpequq_p(__CR6_LT_REV, __a, __b); @@ -16515,39 +17090,92 @@ static __inline__ int __ATTRS_o_ai vec_any_ne(vector unsigned __int128 __a, /* vec_any_nge */ -static __inline__ int __attribute__((__always_inline__)) -vec_any_nge(vector float __a, vector float __b) { +static __inline__ int __ATTRS_o_ai vec_any_nge(vector float __a, + vector float __b) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpgesp_p(__CR6_LT_REV, __a, __b); +#else return __builtin_altivec_vcmpgefp_p(__CR6_LT_REV, __a, __b); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_any_nge(vector double __a, + vector double __b) { + return __builtin_vsx_xvcmpgedp_p(__CR6_LT_REV, __a, __b); +} +#endif + /* vec_any_ngt */ -static __inline__ int __attribute__((__always_inline__)) -vec_any_ngt(vector float __a, vector float __b) { +static __inline__ int __ATTRS_o_ai vec_any_ngt(vector float __a, + vector float __b) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpgtsp_p(__CR6_LT_REV, __a, __b); +#else return __builtin_altivec_vcmpgtfp_p(__CR6_LT_REV, __a, __b); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_any_ngt(vector double __a, + vector double __b) { + return __builtin_vsx_xvcmpgtdp_p(__CR6_LT_REV, __a, __b); +} +#endif + /* vec_any_nle */ -static __inline__ int __attribute__((__always_inline__)) -vec_any_nle(vector float __a, vector float __b) { +static __inline__ int __ATTRS_o_ai vec_any_nle(vector float __a, + vector float __b) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpgesp_p(__CR6_LT_REV, __b, __a); +#else return __builtin_altivec_vcmpgefp_p(__CR6_LT_REV, __b, __a); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_any_nle(vector double __a, + vector double __b) { + return __builtin_vsx_xvcmpgedp_p(__CR6_LT_REV, __b, __a); +} +#endif + /* vec_any_nlt */ -static __inline__ int __attribute__((__always_inline__)) -vec_any_nlt(vector float __a, vector float __b) { +static __inline__ int __ATTRS_o_ai vec_any_nlt(vector float __a, + vector float __b) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpgtsp_p(__CR6_LT_REV, __b, __a); +#else return __builtin_altivec_vcmpgtfp_p(__CR6_LT_REV, __b, __a); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_any_nlt(vector double __a, + vector double __b) { + return __builtin_vsx_xvcmpgtdp_p(__CR6_LT_REV, __b, __a); +} +#endif + /* vec_any_numeric */ -static __inline__ int __attribute__((__always_inline__)) -vec_any_numeric(vector float __a) { +static __inline__ int __ATTRS_o_ai vec_any_numeric(vector float __a) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpeqsp_p(__CR6_EQ_REV, __a, __a); +#else return __builtin_altivec_vcmpeqfp_p(__CR6_EQ_REV, __a, __a); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_any_numeric(vector double __a) { + return __builtin_vsx_xvcmpeqdp_p(__CR6_EQ_REV, __a, __a); +} +#endif + /* vec_any_out */ static __inline__ int __attribute__((__always_inline__)) @@ -16696,6 +17324,16 @@ vec_vgbbd(vector unsigned char __a) { return __builtin_altivec_vgbbd(__a); } +static __inline__ vector signed long long __ATTRS_o_ai +vec_gbb(vector signed long long __a) { + return __builtin_altivec_vgbbd((vector unsigned char)__a); +} + +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_gbb(vector unsigned long long __a) { + return __builtin_altivec_vgbbd((vector unsigned char)__a); +} + static __inline__ vector long long __ATTRS_o_ai vec_vbpermq(vector signed char __a, vector signed char __b) { return __builtin_altivec_vbpermq((vector unsigned char)__a, @@ -16707,7 +17345,7 @@ vec_vbpermq(vector unsigned char __a, vector unsigned char __b) { return __builtin_altivec_vbpermq(__a, __b); } -#ifdef __powerpc64__ +#if defined(__powerpc64__) && defined(__SIZEOF_INT128__) static __inline__ vector unsigned long long __attribute__((__always_inline__)) vec_bperm(vector unsigned __int128 __a, vector unsigned char __b) { return __builtin_altivec_vbpermq((vector unsigned char)__a, @@ -16882,7 +17520,8 @@ vec_revb(vector double __a) { } #endif /* End __VSX__ */ -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_revb(vector signed __int128 __a) { vector unsigned char __indices = @@ -16904,6 +17543,8 @@ vec_revb(vector unsigned __int128 __a) { /* vec_xl */ +#define vec_xld2 vec_xl +#define vec_xlw4 vec_xl typedef vector signed char unaligned_vec_schar __attribute__((aligned(1))); typedef vector unsigned char unaligned_vec_uchar __attribute__((aligned(1))); typedef vector signed short unaligned_vec_sshort __attribute__((aligned(1))); @@ -16912,41 +17553,41 @@ typedef vector signed int unaligned_vec_sint __attribute__((aligned(1))); typedef vector unsigned int unaligned_vec_uint __attribute__((aligned(1))); typedef vector float unaligned_vec_float __attribute__((aligned(1))); -static inline __ATTRS_o_ai vector signed char vec_xl(signed long long __offset, +static inline __ATTRS_o_ai vector signed char vec_xl(ptrdiff_t __offset, const signed char *__ptr) { return *(unaligned_vec_schar *)(__ptr + __offset); } static inline __ATTRS_o_ai vector unsigned char -vec_xl(signed long long __offset, const unsigned char *__ptr) { +vec_xl(ptrdiff_t __offset, const unsigned char *__ptr) { return *(unaligned_vec_uchar*)(__ptr + __offset); } -static inline __ATTRS_o_ai vector signed short vec_xl(signed long long __offset, - const signed short *__ptr) { +static inline __ATTRS_o_ai vector signed short +vec_xl(ptrdiff_t __offset, const signed short *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_sshort *)__addr; } static inline __ATTRS_o_ai vector unsigned short -vec_xl(signed long long __offset, const unsigned short *__ptr) { +vec_xl(ptrdiff_t __offset, const unsigned short *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_ushort *)__addr; } -static inline __ATTRS_o_ai vector signed int vec_xl(signed long long __offset, +static inline __ATTRS_o_ai vector signed int vec_xl(ptrdiff_t __offset, const signed int *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_sint *)__addr; } -static inline __ATTRS_o_ai vector unsigned int vec_xl(signed long long __offset, - const unsigned int *__ptr) { +static inline __ATTRS_o_ai vector unsigned int +vec_xl(ptrdiff_t __offset, const unsigned int *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_uint *)__addr; } -static inline __ATTRS_o_ai vector float vec_xl(signed long long __offset, +static inline __ATTRS_o_ai vector float vec_xl(ptrdiff_t __offset, const float *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_float *)__addr; @@ -16958,36 +17599,37 @@ typedef vector unsigned long long unaligned_vec_ull __attribute__((aligned(1))); typedef vector double unaligned_vec_double __attribute__((aligned(1))); static inline __ATTRS_o_ai vector signed long long -vec_xl(signed long long __offset, const signed long long *__ptr) { +vec_xl(ptrdiff_t __offset, const signed long long *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_sll *)__addr; } static inline __ATTRS_o_ai vector unsigned long long -vec_xl(signed long long __offset, const unsigned long long *__ptr) { +vec_xl(ptrdiff_t __offset, const unsigned long long *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_ull *)__addr; } -static inline __ATTRS_o_ai vector double vec_xl(signed long long __offset, +static inline __ATTRS_o_ai vector double vec_xl(ptrdiff_t __offset, const double *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_double *)__addr; } #endif -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) typedef vector signed __int128 unaligned_vec_si128 __attribute__((aligned(1))); typedef vector unsigned __int128 unaligned_vec_ui128 __attribute__((aligned(1))); static inline __ATTRS_o_ai vector signed __int128 -vec_xl(signed long long __offset, const signed __int128 *__ptr) { +vec_xl(ptrdiff_t __offset, const signed __int128 *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_si128 *)__addr; } static inline __ATTRS_o_ai vector unsigned __int128 -vec_xl(signed long long __offset, const unsigned __int128 *__ptr) { +vec_xl(ptrdiff_t __offset, const unsigned __int128 *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_ui128 *)__addr; } @@ -16997,27 +17639,27 @@ vec_xl(signed long long __offset, const unsigned __int128 *__ptr) { #ifdef __LITTLE_ENDIAN__ static __inline__ vector signed char __ATTRS_o_ai -vec_xl_be(signed long long __offset, const signed char *__ptr) { +vec_xl_be(ptrdiff_t __offset, const signed char *__ptr) { vector signed char __vec = (vector signed char)__builtin_vsx_lxvd2x_be(__offset, __ptr); return __builtin_shufflevector(__vec, __vec, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_xl_be(signed long long __offset, const unsigned char *__ptr) { +vec_xl_be(ptrdiff_t __offset, const unsigned char *__ptr) { vector unsigned char __vec = (vector unsigned char)__builtin_vsx_lxvd2x_be(__offset, __ptr); return __builtin_shufflevector(__vec, __vec, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8); } -static __inline__ vector signed short __ATTRS_o_ai -vec_xl_be(signed long long __offset, const signed short *__ptr) { +static __inline__ vector signed short __ATTRS_o_ai +vec_xl_be(ptrdiff_t __offset, const signed short *__ptr) { vector signed short __vec = (vector signed short)__builtin_vsx_lxvd2x_be(__offset, __ptr); return __builtin_shufflevector(__vec, __vec, 3, 2, 1, 0, 7, 6, 5, 4); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_xl_be(signed long long __offset, const unsigned short *__ptr) { +vec_xl_be(ptrdiff_t __offset, const unsigned short *__ptr) { vector unsigned short __vec = (vector unsigned short)__builtin_vsx_lxvd2x_be(__offset, __ptr); return __builtin_shufflevector(__vec, __vec, 3, 2, 1, 0, 7, 6, 5, 4); } @@ -17054,7 +17696,8 @@ vec_xl_be(signed long long __offset, const double *__ptr) { } #endif -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_xl_be(signed long long __offset, const signed __int128 *__ptr) { return vec_xl(__offset, __ptr); @@ -17069,98 +17712,153 @@ vec_xl_be(signed long long __offset, const unsigned __int128 *__ptr) { #define vec_xl_be vec_xl #endif -#if defined(__POWER10_VECTOR__) && defined(__VSX__) +#if defined(__POWER10_VECTOR__) && defined(__VSX__) && \ + defined(__SIZEOF_INT128__) /* vect_xl_sext */ static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_sext(signed long long __offset, const signed char *__pointer) { +vec_xl_sext(ptrdiff_t __offset, const signed char *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_sext(signed long long __offset, const signed short *__pointer) { +vec_xl_sext(ptrdiff_t __offset, const signed short *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_sext(signed long long __offset, const signed int *__pointer) { +vec_xl_sext(ptrdiff_t __offset, const signed int *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_sext(signed long long __offset, const signed long long *__pointer) { +vec_xl_sext(ptrdiff_t __offset, const signed long long *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } /* vec_xl_zext */ static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_zext(signed long long __offset, const unsigned char *__pointer) { +vec_xl_zext(ptrdiff_t __offset, const unsigned char *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_zext(signed long long __offset, const unsigned short *__pointer) { +vec_xl_zext(ptrdiff_t __offset, const unsigned short *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_zext(signed long long __offset, const unsigned int *__pointer) { +vec_xl_zext(ptrdiff_t __offset, const unsigned int *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_zext(signed long long __offset, const unsigned long long *__pointer) { +vec_xl_zext(ptrdiff_t __offset, const unsigned long long *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } #endif +/* vec_xlds */ +#ifdef __VSX__ +static __inline__ vector signed long long __ATTRS_o_ai +vec_xlds(ptrdiff_t __offset, const signed long long *__ptr) { + signed long long *__addr = (signed long long*)((signed char *)__ptr + __offset); + return (vector signed long long) *__addr; +} + +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_xlds(ptrdiff_t __offset, const unsigned long long *__ptr) { + unsigned long long *__addr = (unsigned long long *)((signed char *)__ptr + __offset); + return (unaligned_vec_ull) *__addr; +} + +static __inline__ vector double __ATTRS_o_ai vec_xlds(ptrdiff_t __offset, + const double *__ptr) { + double *__addr = (double*)((signed char *)__ptr + __offset); + return (unaligned_vec_double) *__addr; +} + +/* vec_load_splats */ +static __inline__ vector signed int __ATTRS_o_ai +vec_load_splats(signed long long __offset, const signed int *__ptr) { + signed int *__addr = (signed int*)((signed char *)__ptr + __offset); + return (vector signed int)*__addr; +} + +static __inline__ vector signed int __ATTRS_o_ai +vec_load_splats(unsigned long long __offset, const signed int *__ptr) { + signed int *__addr = (signed int*)((signed char *)__ptr + __offset); + return (vector signed int)*__addr; +} + +static __inline__ vector unsigned int __ATTRS_o_ai +vec_load_splats(signed long long __offset, const unsigned int *__ptr) { + unsigned int *__addr = (unsigned int*)((signed char *)__ptr + __offset); + return (vector unsigned int)*__addr; +} + +static __inline__ vector unsigned int __ATTRS_o_ai +vec_load_splats(unsigned long long __offset, const unsigned int *__ptr) { + unsigned int *__addr = (unsigned int*)((signed char *)__ptr + __offset); + return (vector unsigned int)*__addr; +} + +static __inline__ vector float __ATTRS_o_ai +vec_load_splats(signed long long __offset, const float *__ptr) { + float *__addr = (float*)((signed char *)__ptr + __offset); + return (vector float)*__addr; +} + +static __inline__ vector float __ATTRS_o_ai +vec_load_splats(unsigned long long __offset, const float *__ptr) { + float *__addr = (float*)((signed char *)__ptr + __offset); + return (vector float)*__addr; +} +#endif + /* vec_xst */ -static inline __ATTRS_o_ai void vec_xst(vector signed char __vec, - signed long long __offset, - signed char *__ptr) { +#define vec_xstd2 vec_xst +#define vec_xstw4 vec_xst +static inline __ATTRS_o_ai void +vec_xst(vector signed char __vec, ptrdiff_t __offset, signed char *__ptr) { *(unaligned_vec_schar *)(__ptr + __offset) = __vec; } -static inline __ATTRS_o_ai void vec_xst(vector unsigned char __vec, - signed long long __offset, - unsigned char *__ptr) { +static inline __ATTRS_o_ai void +vec_xst(vector unsigned char __vec, ptrdiff_t __offset, unsigned char *__ptr) { *(unaligned_vec_uchar *)(__ptr + __offset) = __vec; } -static inline __ATTRS_o_ai void vec_xst(vector signed short __vec, - signed long long __offset, - signed short *__ptr) { +static inline __ATTRS_o_ai void +vec_xst(vector signed short __vec, ptrdiff_t __offset, signed short *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_sshort *)__addr = __vec; } static inline __ATTRS_o_ai void vec_xst(vector unsigned short __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned short *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_ushort *)__addr = __vec; } static inline __ATTRS_o_ai void vec_xst(vector signed int __vec, - signed long long __offset, - signed int *__ptr) { + ptrdiff_t __offset, signed int *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_sint *)__addr = __vec; } -static inline __ATTRS_o_ai void vec_xst(vector unsigned int __vec, - signed long long __offset, - unsigned int *__ptr) { +static inline __ATTRS_o_ai void +vec_xst(vector unsigned int __vec, ptrdiff_t __offset, unsigned int *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_uint *)__addr = __vec; } -static inline __ATTRS_o_ai void vec_xst(vector float __vec, - signed long long __offset, +static inline __ATTRS_o_ai void vec_xst(vector float __vec, ptrdiff_t __offset, float *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_float *)__addr = __vec; @@ -17168,37 +17866,37 @@ static inline __ATTRS_o_ai void vec_xst(vector float __vec, #ifdef __VSX__ static inline __ATTRS_o_ai void vec_xst(vector signed long long __vec, - signed long long __offset, + ptrdiff_t __offset, signed long long *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_sll *)__addr = __vec; } static inline __ATTRS_o_ai void vec_xst(vector unsigned long long __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned long long *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_ull *)__addr = __vec; } -static inline __ATTRS_o_ai void vec_xst(vector double __vec, - signed long long __offset, +static inline __ATTRS_o_ai void vec_xst(vector double __vec, ptrdiff_t __offset, double *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_double *)__addr = __vec; } #endif -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) static inline __ATTRS_o_ai void vec_xst(vector signed __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, signed __int128 *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_si128 *)__addr = __vec; } static inline __ATTRS_o_ai void vec_xst(vector unsigned __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned __int128 *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_ui128 *)__addr = __vec; @@ -17207,51 +17905,52 @@ static inline __ATTRS_o_ai void vec_xst(vector unsigned __int128 __vec, /* vec_xst_trunc */ -#if defined(__POWER10_VECTOR__) && defined(__VSX__) +#if defined(__POWER10_VECTOR__) && defined(__VSX__) && \ + defined(__SIZEOF_INT128__) static inline __ATTRS_o_ai void vec_xst_trunc(vector signed __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, signed char *__ptr) { *(__ptr + __offset) = (signed char)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector unsigned __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned char *__ptr) { *(__ptr + __offset) = (unsigned char)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector signed __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, signed short *__ptr) { *(__ptr + __offset) = (signed short)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector unsigned __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned short *__ptr) { *(__ptr + __offset) = (unsigned short)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector signed __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, signed int *__ptr) { *(__ptr + __offset) = (signed int)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector unsigned __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned int *__ptr) { *(__ptr + __offset) = (unsigned int)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector signed __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, signed long long *__ptr) { *(__ptr + __offset) = (signed long long)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector unsigned __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned long long *__ptr) { *(__ptr + __offset) = (unsigned long long)__vec[0]; } @@ -17336,7 +18035,8 @@ static __inline__ void __ATTRS_o_ai vec_xst_be(vector double __vec, } #endif -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) static __inline__ void __ATTRS_o_ai vec_xst_be(vector signed __int128 __vec, signed long long __offset, signed __int128 *__ptr) { @@ -17375,7 +18075,7 @@ static vector double __ATTRS_o_ai vec_neg(vector double __a) { #endif -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __VSX__ static vector long long __ATTRS_o_ai vec_neg(vector long long __a) { return -__a; } @@ -17404,7 +18104,7 @@ static vector double __ATTRS_o_ai vec_nabs(vector double __a) { #endif -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ static vector long long __ATTRS_o_ai vec_nabs(vector long long __a) { return __builtin_altivec_vminsd(__a, -__a); } @@ -17422,6 +18122,18 @@ static vector signed char __ATTRS_o_ai vec_nabs(vector signed char __a) { return __builtin_altivec_vminsb(__a, -__a); } +static vector float __ATTRS_o_ai vec_recipdiv(vector float __a, + vector float __b) { + return __builtin_ppc_recipdivf(__a, __b); +} + +#ifdef __VSX__ +static vector double __ATTRS_o_ai vec_recipdiv(vector double __a, + vector double __b) { + return __builtin_ppc_recipdivd(__a, __b); +} +#endif + #ifdef __POWER10_VECTOR__ /* vec_extractm */ @@ -17446,10 +18158,12 @@ vec_extractm(vector unsigned long long __a) { return __builtin_altivec_vextractdm(__a); } +#ifdef __SIZEOF_INT128__ static __inline__ unsigned int __ATTRS_o_ai vec_extractm(vector unsigned __int128 __a) { return __builtin_altivec_vextractqm(__a); } +#endif /* vec_expandm */ @@ -17473,10 +18187,12 @@ vec_expandm(vector unsigned long long __a) { return __builtin_altivec_vexpanddm(__a); } +#ifdef __SIZEOF_INT128__ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_expandm(vector unsigned __int128 __a) { return __builtin_altivec_vexpandqm(__a); } +#endif /* vec_cntm */ @@ -17512,10 +18228,12 @@ vec_gendm(unsigned long long __bm) { return __builtin_altivec_mtvsrdm(__bm); } +#ifdef __SIZEOF_INT128__ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_genqm(unsigned long long __bm) { return __builtin_altivec_mtvsrqm(__bm); } +#endif /* vec_pdep */ @@ -17544,6 +18262,7 @@ vec_cfuge(vector unsigned long long __a, vector unsigned long long __b) { /* vec_ternarylogic */ #ifdef __VSX__ +#ifdef __SIZEOF_INT128__ #define vec_ternarylogic(__a, __b, __c, __imm) \ _Generic((__a), vector unsigned char \ : __builtin_vsx_xxeval((vector unsigned long long)(__a), \ @@ -17565,6 +18284,25 @@ vec_cfuge(vector unsigned long long __a, vector unsigned long long __b) { : __builtin_vsx_xxeval((vector unsigned long long)(__a), \ (vector unsigned long long)(__b), \ (vector unsigned long long)(__c), (__imm))) +#else +#define vec_ternarylogic(__a, __b, __c, __imm) \ + _Generic((__a), vector unsigned char \ + : __builtin_vsx_xxeval((vector unsigned long long)(__a), \ + (vector unsigned long long)(__b), \ + (vector unsigned long long)(__c), (__imm)), \ + vector unsigned short \ + : __builtin_vsx_xxeval((vector unsigned long long)(__a), \ + (vector unsigned long long)(__b), \ + (vector unsigned long long)(__c), (__imm)), \ + vector unsigned int \ + : __builtin_vsx_xxeval((vector unsigned long long)(__a), \ + (vector unsigned long long)(__b), \ + (vector unsigned long long)(__c), (__imm)), \ + vector unsigned long long \ + : __builtin_vsx_xxeval((vector unsigned long long)(__a), \ + (vector unsigned long long)(__b), \ + (vector unsigned long long)(__c), (__imm))) +#endif /* __SIZEOF_INT128__ */ #endif /* __VSX__ */ /* vec_genpcvm */ @@ -17657,6 +18395,7 @@ vec_mod(vector unsigned long long __a, vector unsigned long long __b) { return __a % __b; } +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_mod(vector signed __int128 __a, vector signed __int128 __b) { return __a % __b; @@ -17666,6 +18405,7 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_mod(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __a % __b; } +#endif /* vec_sldbi */ @@ -18188,6 +18928,7 @@ static __inline__ int __ATTRS_o_ai vec_strir_p(vector signed short __a) { /* vs[l | r | ra] */ +#ifdef __SIZEOF_INT128__ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_sl(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __a << (__b % (vector unsigned __int128)(sizeof(unsigned __int128) * @@ -18232,6 +18973,7 @@ vec_sra(vector signed __int128 __a, vector unsigned __int128 __b) { __CHAR_BIT__)); } +#endif /* __SIZEOF_INT128__ */ #endif /* __POWER10_VECTOR__ */ #undef __ATTRS_o_ai diff --git a/lib/include/amxintrin.h b/lib/include/amxintrin.h index 823c7ca1f0..ec601a58e7 100644 --- a/lib/include/amxintrin.h +++ b/lib/include/amxintrin.h @@ -15,8 +15,13 @@ #define __AMXINTRIN_H #ifdef __x86_64__ +/* Define the default attributes for the functions in this file. */ #define __DEFAULT_FN_ATTRS_TILE \ __attribute__((__always_inline__, __nodebug__, __target__("amx-tile"))) +#define __DEFAULT_FN_ATTRS_INT8 \ + __attribute__((__always_inline__, __nodebug__, __target__("amx-int8"))) +#define __DEFAULT_FN_ATTRS_BF16 \ + __attribute__((__always_inline__, __nodebug__, __target__("amx-bf16"))) /// Load tile configuration from a 64-byte memory location specified by /// "mem_addr". The tile configuration includes the tile type palette, the @@ -25,7 +30,7 @@ /// config and the tile data, and the tiles are zeroed. Any invalid /// configurations will result in #GP fault. /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the LDTILECFG instruction. /// @@ -41,7 +46,7 @@ _tile_loadconfig(const void *__config) { /// palette, the number of bytes per row, and the number of rows. If tiles /// are not configured, all zeroes will be stored to memory. /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the STTILECFG instruction. /// @@ -55,7 +60,7 @@ _tile_storeconfig(void *__config) { /// Release the tile configuration to return to the init state, which /// releases all storage it currently holds. /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TILERELEASE instruction. static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { @@ -66,7 +71,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// destination tile "dst" using the tile configuration previously configured /// via "_tile_loadconfig". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TILELOADD instruction. /// @@ -86,7 +91,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// that the data will likely not be reused in the near future and the data /// caching can be optimized accordingly. /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TILELOADDT1 instruction. /// @@ -104,7 +109,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// "stride" using the tile configuration previously configured via /// "_tile_loadconfig". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TILESTORED instruction. /// @@ -119,7 +124,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// Zero the tile specified by "tdest". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TILEZERO instruction. /// @@ -133,7 +138,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// results. Sum these 4 results with the corresponding 32-bit integer in "dst", /// and store the 32-bit result back to tile "dst". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TDPBSSD instruction. /// @@ -152,7 +157,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// 32-bit results. Sum these 4 results with the corresponding 32-bit integer /// in "dst", and store the 32-bit result back to tile "dst". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TDPBSUD instruction. /// @@ -171,7 +176,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// results. Sum these 4 results with the corresponding 32-bit integer in "dst", /// and store the 32-bit result back to tile "dst". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TDPBUSD instruction. /// @@ -190,7 +195,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// 32-bit results. Sum these 4 results with the corresponding 32-bit integer in /// "dst", and store the 32-bit result back to tile "dst". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TDPBUUD instruction. /// @@ -208,7 +213,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// elements with elements in "dst", and store the 32-bit result back to tile /// "dst". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TDPBF16PS instruction. /// @@ -221,10 +226,12 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { #define _tile_dpbf16ps(dst, src0, src1) \ __builtin_ia32_tdpbf16ps((dst), (src0), (src1)) -#define __DEFAULT_FN_ATTRS_INT8 \ - __attribute__((__always_inline__, __nodebug__, __target__("amx-int8"))) - +/// AMX tile register size can be configured, the maximum size is 16x64=1024 +/// bytes. Since there is no 2D type in llvm IR, we use vector type to +/// represent 2D tile and the fixed size is maximum amx tile register size. typedef int _tile1024i __attribute__((__vector_size__(1024), __aligned__(64))); + +/// This is internal intrinsic. C/C++ user should avoid calling it directly. static __inline__ _tile1024i __DEFAULT_FN_ATTRS_INT8 _tile_loadd_internal(unsigned short m, unsigned short n, const void *base, __SIZE_TYPE__ stride) { @@ -232,12 +239,43 @@ _tile_loadd_internal(unsigned short m, unsigned short n, const void *base, (__SIZE_TYPE__)(stride)); } +/// This is internal intrinsic. C/C++ user should avoid calling it directly. +static __inline__ _tile1024i __DEFAULT_FN_ATTRS_INT8 +_tile_loaddt1_internal(unsigned short m, unsigned short n, const void *base, + __SIZE_TYPE__ stride) { + return __builtin_ia32_tileloaddt164_internal(m, n, base, + (__SIZE_TYPE__)(stride)); +} + +/// This is internal intrinsic. C/C++ user should avoid calling it directly. static __inline__ _tile1024i __DEFAULT_FN_ATTRS_INT8 _tile_dpbssd_internal(unsigned short m, unsigned short n, unsigned short k, _tile1024i dst, _tile1024i src1, _tile1024i src2) { return __builtin_ia32_tdpbssd_internal(m, n, k, dst, src1, src2); } +/// This is internal intrinsic. C/C++ user should avoid calling it directly. +static __inline__ _tile1024i __DEFAULT_FN_ATTRS_INT8 +_tile_dpbsud_internal(unsigned short m, unsigned short n, unsigned short k, + _tile1024i dst, _tile1024i src1, _tile1024i src2) { + return __builtin_ia32_tdpbsud_internal(m, n, k, dst, src1, src2); +} + +/// This is internal intrinsic. C/C++ user should avoid calling it directly. +static __inline__ _tile1024i __DEFAULT_FN_ATTRS_INT8 +_tile_dpbusd_internal(unsigned short m, unsigned short n, unsigned short k, + _tile1024i dst, _tile1024i src1, _tile1024i src2) { + return __builtin_ia32_tdpbusd_internal(m, n, k, dst, src1, src2); +} + +/// This is internal intrinsic. C/C++ user should avoid calling it directly. +static __inline__ _tile1024i __DEFAULT_FN_ATTRS_INT8 +_tile_dpbuud_internal(unsigned short m, unsigned short n, unsigned short k, + _tile1024i dst, _tile1024i src1, _tile1024i src2) { + return __builtin_ia32_tdpbuud_internal(m, n, k, dst, src1, src2); +} + +/// This is internal intrinsic. C/C++ user should avoid calling it directly. static __inline__ void __DEFAULT_FN_ATTRS_INT8 _tile_stored_internal(unsigned short m, unsigned short n, void *base, __SIZE_TYPE__ stride, _tile1024i tile) { @@ -245,34 +283,211 @@ _tile_stored_internal(unsigned short m, unsigned short n, void *base, (__SIZE_TYPE__)(stride), tile); } +/// This is internal intrinsic. C/C++ user should avoid calling it directly. +static __inline__ _tile1024i __DEFAULT_FN_ATTRS_BF16 +_tile_dpbf16ps_internal(unsigned short m, unsigned short n, unsigned short k, + _tile1024i dst, _tile1024i src1, _tile1024i src2) { + return __builtin_ia32_tdpbf16ps_internal(m, n, k, dst, src1, src2); +} + +/// This struct pack the shape and tile data together for user. We suggest +/// initializing the struct as early as possible, because compiler depends +/// on the shape information to do configure. The constant value is preferred +/// for optimization by compiler. typedef struct __tile1024i_str { const unsigned short row; const unsigned short col; _tile1024i tile; } __tile1024i; +/// Load tile rows from memory specifieid by "base" address and "stride" into +/// destination tile "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TILELOADD instruction. +/// +/// \param dst +/// A destination tile. Max size is 1024 Bytes. +/// \param base +/// A pointer to base address. +/// \param stride +/// The stride between the rows' data to be loaded in memory. __DEFAULT_FN_ATTRS_TILE static void __tile_loadd(__tile1024i *dst, const void *base, __SIZE_TYPE__ stride) { dst->tile = _tile_loadd_internal(dst->row, dst->col, base, stride); } -__DEFAULT_FN_ATTRS_INT8 -static void __tile_dpbssd(__tile1024i *dst, __tile1024i src1, - __tile1024i src2) { - dst->tile = _tile_dpbssd_internal(src1.row, src2.col, src1.col, dst->tile, - src1.tile, src2.tile); +/// Load tile rows from memory specifieid by "base" address and "stride" into +/// destination tile "dst". This intrinsic provides a hint to the implementation +/// that the data will likely not be reused in the near future and the data +/// caching can be optimized accordingly. +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TILELOADDT1 instruction. +/// +/// \param dst +/// A destination tile. Max size is 1024 Bytes. +/// \param base +/// A pointer to base address. +/// \param stride +/// The stride between the rows' data to be loaded in memory. +__DEFAULT_FN_ATTRS_TILE +static void __tile_stream_loadd(__tile1024i *dst, const void *base, + __SIZE_TYPE__ stride) { + dst->tile = _tile_loaddt1_internal(dst->row, dst->col, base, stride); } +/// Compute dot-product of bytes in tiles with a source/destination accumulator. +/// Multiply groups of 4 adjacent pairs of signed 8-bit integers in src0 with +/// corresponding signed 8-bit integers in src1, producing 4 intermediate 32-bit +/// results. Sum these 4 results with the corresponding 32-bit integer in "dst", +/// and store the 32-bit result back to tile "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TDPBSSD instruction. +/// +/// \param dst +/// The destination tile. Max size is 1024 Bytes. +/// \param src0 +/// The 1st source tile. Max size is 1024 Bytes. +/// \param src1 +/// The 2nd source tile. Max size is 1024 Bytes. +__DEFAULT_FN_ATTRS_INT8 +static void __tile_dpbssd(__tile1024i *dst, __tile1024i src0, + __tile1024i src1) { + dst->tile = _tile_dpbssd_internal(src0.row, src1.col, src0.col, dst->tile, + src0.tile, src1.tile); +} + +/// Compute dot-product of bytes in tiles with a source/destination accumulator. +/// Multiply groups of 4 adjacent pairs of signed 8-bit integers in src0 with +/// corresponding unsigned 8-bit integers in src1, producing 4 intermediate +/// 32-bit results. Sum these 4 results with the corresponding 32-bit integer +/// in "dst", and store the 32-bit result back to tile "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TDPBSUD instruction. +/// +/// \param dst +/// The destination tile. Max size is 1024 Bytes. +/// \param src0 +/// The 1st source tile. Max size is 1024 Bytes. +/// \param src1 +/// The 2nd source tile. Max size is 1024 Bytes. +__DEFAULT_FN_ATTRS_INT8 +static void __tile_dpbsud(__tile1024i *dst, __tile1024i src0, + __tile1024i src1) { + dst->tile = _tile_dpbsud_internal(src0.row, src1.col, src0.col, dst->tile, + src0.tile, src1.tile); +} + +/// Compute dot-product of bytes in tiles with a source/destination accumulator. +/// Multiply groups of 4 adjacent pairs of unsigned 8-bit integers in src0 with +/// corresponding signed 8-bit integers in src1, producing 4 intermediate 32-bit +/// results. Sum these 4 results with the corresponding 32-bit integer in "dst", +/// and store the 32-bit result back to tile "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TDPBUSD instruction. +/// +/// \param dst +/// The destination tile. Max size is 1024 Bytes. +/// \param src0 +/// The 1st source tile. Max size is 1024 Bytes. +/// \param src1 +/// The 2nd source tile. Max size is 1024 Bytes. +__DEFAULT_FN_ATTRS_INT8 +static void __tile_dpbusd(__tile1024i *dst, __tile1024i src0, + __tile1024i src1) { + dst->tile = _tile_dpbusd_internal(src0.row, src1.col, src0.col, dst->tile, + src0.tile, src1.tile); +} + +/// Compute dot-product of bytes in tiles with a source/destination accumulator. +/// Multiply groups of 4 adjacent pairs of unsigned 8-bit integers in src0 with +/// corresponding unsigned 8-bit integers in src1, producing 4 intermediate +/// 32-bit results. Sum these 4 results with the corresponding 32-bit integer in +/// "dst", and store the 32-bit result back to tile "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TDPBUUD instruction. +/// +/// \param dst +/// The destination tile. Max size is 1024 Bytes. +/// \param src0 +/// The 1st source tile. Max size is 1024 Bytes. +/// \param src1 +/// The 2nd source tile. Max size is 1024 Bytes. +__DEFAULT_FN_ATTRS_INT8 +static void __tile_dpbuud(__tile1024i *dst, __tile1024i src0, + __tile1024i src1) { + dst->tile = _tile_dpbuud_internal(src0.row, src1.col, src0.col, dst->tile, + src0.tile, src1.tile); +} + +/// Store the tile specified by "src" to memory specifieid by "base" address and +/// "stride". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TILESTORED instruction. +/// +/// \param dst +/// A destination tile. Max size is 1024 Bytes. +/// \param base +/// A pointer to base address. +/// \param stride +/// The stride between the rows' data to be stored in memory. __DEFAULT_FN_ATTRS_TILE static void __tile_stored(void *base, __SIZE_TYPE__ stride, __tile1024i src) { _tile_stored_internal(src.row, src.col, base, stride, src.tile); } +/// Zero the tile specified by "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TILEZERO instruction. +/// +/// \param dst +/// The destination tile to be zero. Max size is 1024 Bytes. __DEFAULT_FN_ATTRS_TILE static void __tile_zero(__tile1024i *dst) { dst->tile = __builtin_ia32_tilezero_internal(dst->row, dst->col); } +/// Compute dot-product of BF16 (16-bit) floating-point pairs in tiles src0 and +/// src1, accumulating the intermediate single-precision (32-bit) floating-point +/// elements with elements in "dst", and store the 32-bit result back to tile +/// "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TDPBF16PS instruction. +/// +/// \param dst +/// The destination tile. Max size is 1024 Bytes. +/// \param src0 +/// The 1st source tile. Max size is 1024 Bytes. +/// \param src1 +/// The 2nd source tile. Max size is 1024 Bytes. +__DEFAULT_FN_ATTRS_BF16 +static void __tile_dpbf16ps(__tile1024i *dst, __tile1024i src0, + __tile1024i src1) { + dst->tile = _tile_dpbf16ps_internal(src0.row, src1.col, src0.col, dst->tile, + src0.tile, src1.tile); +} + +#undef __DEFAULT_FN_ATTRS_TILE +#undef __DEFAULT_FN_ATTRS_INT8 +#undef __DEFAULT_FN_ATTRS_BF16 + #endif /* __x86_64__ */ #endif /* __AMXINTRIN_H */ diff --git a/lib/include/arm_acle.h b/lib/include/arm_acle.h index c156d89c1f..45fac248da 100644 --- a/lib/include/arm_acle.h +++ b/lib/include/arm_acle.h @@ -639,6 +639,49 @@ __jcvt(double __a) { } #endif +/* Armv8.5-A FP rounding intrinsics */ +#if __ARM_64BIT_STATE && defined(__ARM_FEATURE_FRINT) +static __inline__ float __attribute__((__always_inline__, __nodebug__)) +__frint32zf(float __a) { + return __builtin_arm_frint32zf(__a); +} + +static __inline__ double __attribute__((__always_inline__, __nodebug__)) +__frint32z(double __a) { + return __builtin_arm_frint32z(__a); +} + +static __inline__ float __attribute__((__always_inline__, __nodebug__)) +__frint64zf(float __a) { + return __builtin_arm_frint64zf(__a); +} + +static __inline__ double __attribute__((__always_inline__, __nodebug__)) +__frint64z(double __a) { + return __builtin_arm_frint64z(__a); +} + +static __inline__ float __attribute__((__always_inline__, __nodebug__)) +__frint32xf(float __a) { + return __builtin_arm_frint32xf(__a); +} + +static __inline__ double __attribute__((__always_inline__, __nodebug__)) +__frint32x(double __a) { + return __builtin_arm_frint32x(__a); +} + +static __inline__ float __attribute__((__always_inline__, __nodebug__)) +__frint64xf(float __a) { + return __builtin_arm_frint64xf(__a); +} + +static __inline__ double __attribute__((__always_inline__, __nodebug__)) +__frint64x(double __a) { + return __builtin_arm_frint64x(__a); +} +#endif + /* Armv8.7-A load/store 64-byte intrinsics */ #if __ARM_64BIT_STATE && defined(__ARM_FEATURE_LS64) typedef struct { @@ -709,6 +752,18 @@ __arm_st64bv0(void *__addr, data512_t __value) { #endif /* __ARM_FEATURE_TME */ +/* Armv8.5-A Random number generation intrinsics */ +#if __ARM_64BIT_STATE && defined(__ARM_FEATURE_RNG) +static __inline__ int __attribute__((__always_inline__, __nodebug__)) +__rndr(uint64_t *__p) { + return __builtin_arm_rndr(__p); +} +static __inline__ int __attribute__((__always_inline__, __nodebug__)) +__rndrrs(uint64_t *__p) { + return __builtin_arm_rndrrs(__p); +} +#endif + #if defined(__cplusplus) } #endif diff --git a/lib/include/arm_neon.h b/lib/include/arm_neon.h index dbb65b8a49..2448870ce4 100644 --- a/lib/include/arm_neon.h +++ b/lib/include/arm_neon.h @@ -2432,6 +2432,96 @@ __ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) { } #endif +#ifdef __LITTLE_ENDIAN__ +__ai poly8x8_t vadd_p8(poly8x8_t __p0, poly8x8_t __p1) { + poly8x8_t __ret; + __ret = (poly8x8_t) __builtin_neon_vadd_v((int8x8_t)__p0, (int8x8_t)__p1, 4); + return __ret; +} +#else +__ai poly8x8_t vadd_p8(poly8x8_t __p0, poly8x8_t __p1) { + poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0); + poly8x8_t __ret; + __ret = (poly8x8_t) __builtin_neon_vadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +__ai poly64x1_t vadd_p64(poly64x1_t __p0, poly64x1_t __p1) { + poly64x1_t __ret; + __ret = (poly64x1_t) __builtin_neon_vadd_v((int8x8_t)__p0, (int8x8_t)__p1, 6); + return __ret; +} +#ifdef __LITTLE_ENDIAN__ +__ai poly16x4_t vadd_p16(poly16x4_t __p0, poly16x4_t __p1) { + poly16x4_t __ret; + __ret = (poly16x4_t) __builtin_neon_vadd_v((int8x8_t)__p0, (int8x8_t)__p1, 5); + return __ret; +} +#else +__ai poly16x4_t vadd_p16(poly16x4_t __p0, poly16x4_t __p1) { + poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + poly16x4_t __ret; + __ret = (poly16x4_t) __builtin_neon_vadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 5); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai poly8x16_t vaddq_p8(poly8x16_t __p0, poly8x16_t __p1) { + poly8x16_t __ret; + __ret = (poly8x16_t) __builtin_neon_vaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 36); + return __ret; +} +#else +__ai poly8x16_t vaddq_p8(poly8x16_t __p0, poly8x16_t __p1) { + poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + poly8x16_t __ret; + __ret = (poly8x16_t) __builtin_neon_vaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36); + __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai poly64x2_t vaddq_p64(poly64x2_t __p0, poly64x2_t __p1) { + poly64x2_t __ret; + __ret = (poly64x2_t) __builtin_neon_vaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 38); + return __ret; +} +#else +__ai poly64x2_t vaddq_p64(poly64x2_t __p0, poly64x2_t __p1) { + poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + poly64x2_t __ret; + __ret = (poly64x2_t) __builtin_neon_vaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 38); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai poly16x8_t vaddq_p16(poly16x8_t __p0, poly16x8_t __p1) { + poly16x8_t __ret; + __ret = (poly16x8_t) __builtin_neon_vaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 37); + return __ret; +} +#else +__ai poly16x8_t vaddq_p16(poly16x8_t __p0, poly16x8_t __p1) { + poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0); + poly16x8_t __ret; + __ret = (poly16x8_t) __builtin_neon_vaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 37); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + #ifdef __LITTLE_ENDIAN__ __ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) { uint16x4_t __ret; @@ -4552,6 +4642,54 @@ __ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) { } #endif +#ifdef __LITTLE_ENDIAN__ +__ai int8x16_t vclsq_u8(uint8x16_t __p0) { + int8x16_t __ret; + __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 32); + return __ret; +} +#else +__ai int8x16_t vclsq_u8(uint8x16_t __p0) { + uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __ret; + __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 32); + __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int32x4_t vclsq_u32(uint32x4_t __p0) { + int32x4_t __ret; + __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 34); + return __ret; +} +#else +__ai int32x4_t vclsq_u32(uint32x4_t __p0) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + int32x4_t __ret; + __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 34); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int16x8_t vclsq_u16(uint16x8_t __p0) { + int16x8_t __ret; + __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 33); + return __ret; +} +#else +__ai int16x8_t vclsq_u16(uint16x8_t __p0) { + uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __ret; + __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 33); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + #ifdef __LITTLE_ENDIAN__ __ai int8x16_t vclsq_s8(int8x16_t __p0) { int8x16_t __ret; @@ -4600,6 +4738,54 @@ __ai int16x8_t vclsq_s16(int16x8_t __p0) { } #endif +#ifdef __LITTLE_ENDIAN__ +__ai int8x8_t vcls_u8(uint8x8_t __p0) { + int8x8_t __ret; + __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__p0, 0); + return __ret; +} +#else +__ai int8x8_t vcls_u8(uint8x8_t __p0) { + uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + int8x8_t __ret; + __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 0); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int32x2_t vcls_u32(uint32x2_t __p0) { + int32x2_t __ret; + __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__p0, 2); + return __ret; +} +#else +__ai int32x2_t vcls_u32(uint32x2_t __p0) { + uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + int32x2_t __ret; + __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 2); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int16x4_t vcls_u16(uint16x4_t __p0) { + int16x4_t __ret; + __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__p0, 1); + return __ret; +} +#else +__ai int16x4_t vcls_u16(uint16x4_t __p0) { + uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + int16x4_t __ret; + __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 1); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + #ifdef __LITTLE_ENDIAN__ __ai int8x8_t vcls_s8(int8x8_t __p0) { int8x8_t __ret; @@ -34889,7 +35075,7 @@ __ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) { #endif #endif -#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_CRYPTO) +#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_AES) #ifdef __LITTLE_ENDIAN__ __ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) { uint8x16_t __ret; @@ -34956,168 +35142,6 @@ __ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) { } #endif -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32(__p0, __p1, __p2); - return __ret; -} -#else -__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32(__rev0, __p1, __rev2); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -__ai uint32_t vsha1h_u32(uint32_t __p0) { - uint32_t __ret; - __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0); - return __ret; -} -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32(__p0, __p1, __p2); - return __ret; -} -#else -__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32(__rev0, __p1, __rev2); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32(__p0, __p1, __p2); - return __ret; -} -#else -__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32(__rev0, __p1, __rev2); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); - return __ret; -} -#else -__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__p0, (int8x16_t)__p1, 50); - return __ret; -} -#else -__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); - return __ret; -} -#else -__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); - return __ret; -} -#else -__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__p0, (int8x16_t)__p1, 50); - return __ret; -} -#else -__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); - return __ret; -} -#else -__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - #endif #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_DIRECTED_ROUNDING) #ifdef __LITTLE_ENDIAN__ @@ -35683,6 +35707,759 @@ __ai float16x4_t vminnm_f16(float16x4_t __p0, float16x4_t __p1) { } #endif +#endif +#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_SHA2) +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32(__p0, __p1, __p2); + return __ret; +} +#else +__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32(__rev0, __p1, __rev2); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +__ai uint32_t vsha1h_u32(uint32_t __p0) { + uint32_t __ret; + __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0); + return __ret; +} +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32(__p0, __p1, __p2); + return __ret; +} +#else +__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32(__rev0, __p1, __rev2); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32(__p0, __p1, __p2); + return __ret; +} +#else +__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32(__rev0, __p1, __rev2); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__p0, (int8x16_t)__p1, 50); + return __ret; +} +#else +__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__p0, (int8x16_t)__p1, 50); + return __ret; +} +#else +__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#endif +#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_SHA3) && defined(__aarch64__) +#ifdef __LITTLE_ENDIAN__ +__ai uint8x16_t vbcaxq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) { + uint8x16_t __ret; + __ret = (uint8x16_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48); + return __ret; +} +#else +__ai uint8x16_t vbcaxq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) { + uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + uint8x16_t __ret; + __ret = (uint8x16_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48); + __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vbcaxq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vbcaxq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t vbcaxq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51); + return __ret; +} +#else +__ai uint64x2_t vbcaxq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint16x8_t vbcaxq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) { + uint16x8_t __ret; + __ret = (uint16x8_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49); + return __ret; +} +#else +__ai uint16x8_t vbcaxq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) { + uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0); + uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0); + uint16x8_t __ret; + __ret = (uint16x8_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int8x16_t vbcaxq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) { + int8x16_t __ret; + __ret = (int8x16_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32); + return __ret; +} +#else +__ai int8x16_t vbcaxq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) { + int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __ret; + __ret = (int8x16_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32); + __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int32x4_t vbcaxq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) { + int32x4_t __ret; + __ret = (int32x4_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34); + return __ret; +} +#else +__ai int32x4_t vbcaxq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) { + int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + int32x4_t __ret; + __ret = (int32x4_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int64x2_t vbcaxq_s64(int64x2_t __p0, int64x2_t __p1, int64x2_t __p2) { + int64x2_t __ret; + __ret = (int64x2_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35); + return __ret; +} +#else +__ai int64x2_t vbcaxq_s64(int64x2_t __p0, int64x2_t __p1, int64x2_t __p2) { + int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + int64x2_t __ret; + __ret = (int64x2_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int16x8_t vbcaxq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) { + int16x8_t __ret; + __ret = (int16x8_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33); + return __ret; +} +#else +__ai int16x8_t vbcaxq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) { + int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __ret; + __ret = (int16x8_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint8x16_t veor3q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) { + uint8x16_t __ret; + __ret = (uint8x16_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48); + return __ret; +} +#else +__ai uint8x16_t veor3q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) { + uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + uint8x16_t __ret; + __ret = (uint8x16_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48); + __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t veor3q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t veor3q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t veor3q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51); + return __ret; +} +#else +__ai uint64x2_t veor3q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint16x8_t veor3q_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) { + uint16x8_t __ret; + __ret = (uint16x8_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49); + return __ret; +} +#else +__ai uint16x8_t veor3q_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) { + uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0); + uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0); + uint16x8_t __ret; + __ret = (uint16x8_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int8x16_t veor3q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) { + int8x16_t __ret; + __ret = (int8x16_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32); + return __ret; +} +#else +__ai int8x16_t veor3q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) { + int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __ret; + __ret = (int8x16_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32); + __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int32x4_t veor3q_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) { + int32x4_t __ret; + __ret = (int32x4_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34); + return __ret; +} +#else +__ai int32x4_t veor3q_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) { + int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + int32x4_t __ret; + __ret = (int32x4_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int64x2_t veor3q_s64(int64x2_t __p0, int64x2_t __p1, int64x2_t __p2) { + int64x2_t __ret; + __ret = (int64x2_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35); + return __ret; +} +#else +__ai int64x2_t veor3q_s64(int64x2_t __p0, int64x2_t __p1, int64x2_t __p2) { + int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + int64x2_t __ret; + __ret = (int64x2_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int16x8_t veor3q_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) { + int16x8_t __ret; + __ret = (int16x8_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33); + return __ret; +} +#else +__ai int16x8_t veor3q_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) { + int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __ret; + __ret = (int16x8_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t vrax1q_u64(uint64x2_t __p0, uint64x2_t __p1) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vrax1q_v((int8x16_t)__p0, (int8x16_t)__p1, 51); + return __ret; +} +#else +__ai uint64x2_t vrax1q_u64(uint64x2_t __p0, uint64x2_t __p1) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vrax1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +#define vxarq_u64(__p0, __p1, __p2) __extension__ ({ \ + uint64x2_t __s0 = __p0; \ + uint64x2_t __s1 = __p1; \ + uint64x2_t __ret; \ + __ret = (uint64x2_t) __builtin_neon_vxarq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \ + __ret; \ +}) +#else +#define vxarq_u64(__p0, __p1, __p2) __extension__ ({ \ + uint64x2_t __s0 = __p0; \ + uint64x2_t __s1 = __p1; \ + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \ + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \ + uint64x2_t __ret; \ + __ret = (uint64x2_t) __builtin_neon_vxarq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \ + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \ + __ret; \ +}) +#endif + +#endif +#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_SHA512) && defined(__aarch64__) +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t vsha512hq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512hq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51); + return __ret; +} +#else +__ai uint64x2_t vsha512hq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512hq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t vsha512h2q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512h2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51); + return __ret; +} +#else +__ai uint64x2_t vsha512h2q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512h2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t vsha512su0q_u64(uint64x2_t __p0, uint64x2_t __p1) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512su0q_v((int8x16_t)__p0, (int8x16_t)__p1, 51); + return __ret; +} +#else +__ai uint64x2_t vsha512su0q_u64(uint64x2_t __p0, uint64x2_t __p1) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t vsha512su1q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512su1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51); + return __ret; +} +#else +__ai uint64x2_t vsha512su1q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#endif +#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_SM3) && defined(__aarch64__) +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsm3partw1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm3partw1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsm3partw1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm3partw1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsm3partw2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm3partw2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsm3partw2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm3partw2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsm3ss1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm3ss1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsm3ss1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm3ss1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +#define vsm3tt1aq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt1aq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 50); \ + __ret; \ +}) +#else +#define vsm3tt1aq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \ + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \ + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt1aq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 50); \ + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \ + __ret; \ +}) +#endif + +#ifdef __LITTLE_ENDIAN__ +#define vsm3tt1bq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt1bq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 50); \ + __ret; \ +}) +#else +#define vsm3tt1bq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \ + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \ + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt1bq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 50); \ + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \ + __ret; \ +}) +#endif + +#ifdef __LITTLE_ENDIAN__ +#define vsm3tt2aq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt2aq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 50); \ + __ret; \ +}) +#else +#define vsm3tt2aq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \ + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \ + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt2aq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 50); \ + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \ + __ret; \ +}) +#endif + +#ifdef __LITTLE_ENDIAN__ +#define vsm3tt2bq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt2bq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 50); \ + __ret; \ +}) +#else +#define vsm3tt2bq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \ + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \ + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt2bq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 50); \ + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \ + __ret; \ +}) +#endif + +#endif +#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_SM4) && defined(__aarch64__) +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsm4eq_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm4eq_v((int8x16_t)__p0, (int8x16_t)__p1, 50); + return __ret; +} +#else +__ai uint32x4_t vsm4eq_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm4eq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsm4ekeyq_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm4ekeyq_v((int8x16_t)__p0, (int8x16_t)__p1, 50); + return __ret; +} +#else +__ai uint32x4_t vsm4ekeyq_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm4ekeyq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + #endif #if __ARM_ARCH >= 8 && defined(__aarch64__) #ifdef __LITTLE_ENDIAN__ @@ -37962,6 +38739,136 @@ __ai float64x1_t vrndx_f64(float64x1_t __p0) { __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10); return __ret; } +#endif +#if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_FRINT) +#ifdef __LITTLE_ENDIAN__ +__ai float32x4_t vrnd32xq_f32(float32x4_t __p0) { + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd32xq_v((int8x16_t)__p0, 41); + return __ret; +} +#else +__ai float32x4_t vrnd32xq_f32(float32x4_t __p0) { + float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd32xq_v((int8x16_t)__rev0, 41); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x2_t vrnd32x_f32(float32x2_t __p0) { + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd32x_v((int8x8_t)__p0, 9); + return __ret; +} +#else +__ai float32x2_t vrnd32x_f32(float32x2_t __p0) { + float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd32x_v((int8x8_t)__rev0, 9); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x4_t vrnd32zq_f32(float32x4_t __p0) { + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd32zq_v((int8x16_t)__p0, 41); + return __ret; +} +#else +__ai float32x4_t vrnd32zq_f32(float32x4_t __p0) { + float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd32zq_v((int8x16_t)__rev0, 41); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x2_t vrnd32z_f32(float32x2_t __p0) { + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd32z_v((int8x8_t)__p0, 9); + return __ret; +} +#else +__ai float32x2_t vrnd32z_f32(float32x2_t __p0) { + float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd32z_v((int8x8_t)__rev0, 9); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x4_t vrnd64xq_f32(float32x4_t __p0) { + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd64xq_v((int8x16_t)__p0, 41); + return __ret; +} +#else +__ai float32x4_t vrnd64xq_f32(float32x4_t __p0) { + float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd64xq_v((int8x16_t)__rev0, 41); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x2_t vrnd64x_f32(float32x2_t __p0) { + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd64x_v((int8x8_t)__p0, 9); + return __ret; +} +#else +__ai float32x2_t vrnd64x_f32(float32x2_t __p0) { + float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd64x_v((int8x8_t)__rev0, 9); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x4_t vrnd64zq_f32(float32x4_t __p0) { + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd64zq_v((int8x16_t)__p0, 41); + return __ret; +} +#else +__ai float32x4_t vrnd64zq_f32(float32x4_t __p0) { + float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd64zq_v((int8x16_t)__rev0, 41); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x2_t vrnd64z_f32(float32x2_t __p0) { + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd64z_v((int8x8_t)__p0, 9); + return __ret; +} +#else +__ai float32x2_t vrnd64z_f32(float32x2_t __p0) { + float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd64z_v((int8x8_t)__rev0, 9); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + #endif #if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_NUMERIC_MAXMIN) #ifdef __LITTLE_ENDIAN__ @@ -42710,7 +43617,7 @@ __ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) #endif #endif -#if defined(__ARM_FEATURE_FP16FML) && defined(__aarch64__) +#if defined(__ARM_FEATURE_FP16_FML) && defined(__aarch64__) #ifdef __LITTLE_ENDIAN__ __ai float32x4_t vfmlalq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) { float32x4_t __ret; @@ -46786,6 +47693,11 @@ __ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) { __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1); return __ret; } +__ai poly128_t vaddq_p128(poly128_t __p0, poly128_t __p1) { + poly128_t __ret; + __ret = (poly128_t) __builtin_neon_vaddq_p128(__p0, __p1); + return __ret; +} #ifdef __LITTLE_ENDIAN__ __ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { uint16x8_t __ret; @@ -67227,7 +68139,7 @@ __ai float32x4_t vcvtq_low_f32_bf16(bfloat16x8_t __p0) { #endif #endif -#if defined(__ARM_FEATURE_FP16FML) && defined(__aarch64__) +#if defined(__ARM_FEATURE_FP16_FML) && defined(__aarch64__) #ifdef __LITTLE_ENDIAN__ #define vfmlalq_lane_high_f16(__p0_869, __p1_869, __p2_869, __p3_869) __extension__ ({ \ float32x4_t __s0_869 = __p0_869; \ diff --git a/lib/include/arm_sve.h b/lib/include/arm_sve.h index 8a03f9da58..909634a562 100644 --- a/lib/include/arm_sve.h +++ b/lib/include/arm_sve.h @@ -132,7 +132,9 @@ enum svprfop }; /* Function attributes */ -#define __aio static inline __attribute__((__always_inline__, __nodebug__, __overloadable__)) +#define __ai static __inline__ __attribute__((__always_inline__, __nodebug__)) + +#define __aio static __inline__ __attribute__((__always_inline__, __nodebug__, __overloadable__)) #define svreinterpret_s8_s8(...) __builtin_sve_reinterpret_s8_s8(__VA_ARGS__) #define svreinterpret_s8_s16(...) __builtin_sve_reinterpret_s8_s16(__VA_ARGS__) @@ -946,3825 +948,7644 @@ __aio svfloat64_t svreinterpret_f64(svfloat64_t op) { return __builtin_sve_reinterpret_f64_f64(op); } -#define svabd_n_f64_m(...) __builtin_sve_svabd_n_f64_m(__VA_ARGS__) -#define svabd_n_f32_m(...) __builtin_sve_svabd_n_f32_m(__VA_ARGS__) -#define svabd_n_f16_m(...) __builtin_sve_svabd_n_f16_m(__VA_ARGS__) -#define svabd_n_f64_x(...) __builtin_sve_svabd_n_f64_x(__VA_ARGS__) -#define svabd_n_f32_x(...) __builtin_sve_svabd_n_f32_x(__VA_ARGS__) -#define svabd_n_f16_x(...) __builtin_sve_svabd_n_f16_x(__VA_ARGS__) -#define svabd_n_f64_z(...) __builtin_sve_svabd_n_f64_z(__VA_ARGS__) -#define svabd_n_f32_z(...) __builtin_sve_svabd_n_f32_z(__VA_ARGS__) -#define svabd_n_f16_z(...) __builtin_sve_svabd_n_f16_z(__VA_ARGS__) -#define svabd_n_s8_m(...) __builtin_sve_svabd_n_s8_m(__VA_ARGS__) -#define svabd_n_s32_m(...) __builtin_sve_svabd_n_s32_m(__VA_ARGS__) -#define svabd_n_s64_m(...) __builtin_sve_svabd_n_s64_m(__VA_ARGS__) -#define svabd_n_s16_m(...) __builtin_sve_svabd_n_s16_m(__VA_ARGS__) -#define svabd_n_s8_x(...) __builtin_sve_svabd_n_s8_x(__VA_ARGS__) -#define svabd_n_s32_x(...) __builtin_sve_svabd_n_s32_x(__VA_ARGS__) -#define svabd_n_s64_x(...) __builtin_sve_svabd_n_s64_x(__VA_ARGS__) -#define svabd_n_s16_x(...) __builtin_sve_svabd_n_s16_x(__VA_ARGS__) -#define svabd_n_s8_z(...) __builtin_sve_svabd_n_s8_z(__VA_ARGS__) -#define svabd_n_s32_z(...) __builtin_sve_svabd_n_s32_z(__VA_ARGS__) -#define svabd_n_s64_z(...) __builtin_sve_svabd_n_s64_z(__VA_ARGS__) -#define svabd_n_s16_z(...) __builtin_sve_svabd_n_s16_z(__VA_ARGS__) -#define svabd_n_u8_m(...) __builtin_sve_svabd_n_u8_m(__VA_ARGS__) -#define svabd_n_u32_m(...) __builtin_sve_svabd_n_u32_m(__VA_ARGS__) -#define svabd_n_u64_m(...) __builtin_sve_svabd_n_u64_m(__VA_ARGS__) -#define svabd_n_u16_m(...) __builtin_sve_svabd_n_u16_m(__VA_ARGS__) -#define svabd_n_u8_x(...) __builtin_sve_svabd_n_u8_x(__VA_ARGS__) -#define svabd_n_u32_x(...) __builtin_sve_svabd_n_u32_x(__VA_ARGS__) -#define svabd_n_u64_x(...) __builtin_sve_svabd_n_u64_x(__VA_ARGS__) -#define svabd_n_u16_x(...) __builtin_sve_svabd_n_u16_x(__VA_ARGS__) -#define svabd_n_u8_z(...) __builtin_sve_svabd_n_u8_z(__VA_ARGS__) -#define svabd_n_u32_z(...) __builtin_sve_svabd_n_u32_z(__VA_ARGS__) -#define svabd_n_u64_z(...) __builtin_sve_svabd_n_u64_z(__VA_ARGS__) -#define svabd_n_u16_z(...) __builtin_sve_svabd_n_u16_z(__VA_ARGS__) -#define svabd_f64_m(...) __builtin_sve_svabd_f64_m(__VA_ARGS__) -#define svabd_f32_m(...) __builtin_sve_svabd_f32_m(__VA_ARGS__) -#define svabd_f16_m(...) __builtin_sve_svabd_f16_m(__VA_ARGS__) -#define svabd_f64_x(...) __builtin_sve_svabd_f64_x(__VA_ARGS__) -#define svabd_f32_x(...) __builtin_sve_svabd_f32_x(__VA_ARGS__) -#define svabd_f16_x(...) __builtin_sve_svabd_f16_x(__VA_ARGS__) -#define svabd_f64_z(...) __builtin_sve_svabd_f64_z(__VA_ARGS__) -#define svabd_f32_z(...) __builtin_sve_svabd_f32_z(__VA_ARGS__) -#define svabd_f16_z(...) __builtin_sve_svabd_f16_z(__VA_ARGS__) -#define svabd_s8_m(...) __builtin_sve_svabd_s8_m(__VA_ARGS__) -#define svabd_s32_m(...) __builtin_sve_svabd_s32_m(__VA_ARGS__) -#define svabd_s64_m(...) __builtin_sve_svabd_s64_m(__VA_ARGS__) -#define svabd_s16_m(...) __builtin_sve_svabd_s16_m(__VA_ARGS__) -#define svabd_s8_x(...) __builtin_sve_svabd_s8_x(__VA_ARGS__) -#define svabd_s32_x(...) __builtin_sve_svabd_s32_x(__VA_ARGS__) -#define svabd_s64_x(...) __builtin_sve_svabd_s64_x(__VA_ARGS__) -#define svabd_s16_x(...) __builtin_sve_svabd_s16_x(__VA_ARGS__) -#define svabd_s8_z(...) __builtin_sve_svabd_s8_z(__VA_ARGS__) -#define svabd_s32_z(...) __builtin_sve_svabd_s32_z(__VA_ARGS__) -#define svabd_s64_z(...) __builtin_sve_svabd_s64_z(__VA_ARGS__) -#define svabd_s16_z(...) __builtin_sve_svabd_s16_z(__VA_ARGS__) -#define svabd_u8_m(...) __builtin_sve_svabd_u8_m(__VA_ARGS__) -#define svabd_u32_m(...) __builtin_sve_svabd_u32_m(__VA_ARGS__) -#define svabd_u64_m(...) __builtin_sve_svabd_u64_m(__VA_ARGS__) -#define svabd_u16_m(...) __builtin_sve_svabd_u16_m(__VA_ARGS__) -#define svabd_u8_x(...) __builtin_sve_svabd_u8_x(__VA_ARGS__) -#define svabd_u32_x(...) __builtin_sve_svabd_u32_x(__VA_ARGS__) -#define svabd_u64_x(...) __builtin_sve_svabd_u64_x(__VA_ARGS__) -#define svabd_u16_x(...) __builtin_sve_svabd_u16_x(__VA_ARGS__) -#define svabd_u8_z(...) __builtin_sve_svabd_u8_z(__VA_ARGS__) -#define svabd_u32_z(...) __builtin_sve_svabd_u32_z(__VA_ARGS__) -#define svabd_u64_z(...) __builtin_sve_svabd_u64_z(__VA_ARGS__) -#define svabd_u16_z(...) __builtin_sve_svabd_u16_z(__VA_ARGS__) -#define svabs_f64_m(...) __builtin_sve_svabs_f64_m(__VA_ARGS__) -#define svabs_f32_m(...) __builtin_sve_svabs_f32_m(__VA_ARGS__) -#define svabs_f16_m(...) __builtin_sve_svabs_f16_m(__VA_ARGS__) -#define svabs_f64_x(...) __builtin_sve_svabs_f64_x(__VA_ARGS__) -#define svabs_f32_x(...) __builtin_sve_svabs_f32_x(__VA_ARGS__) -#define svabs_f16_x(...) __builtin_sve_svabs_f16_x(__VA_ARGS__) -#define svabs_f64_z(...) __builtin_sve_svabs_f64_z(__VA_ARGS__) -#define svabs_f32_z(...) __builtin_sve_svabs_f32_z(__VA_ARGS__) -#define svabs_f16_z(...) __builtin_sve_svabs_f16_z(__VA_ARGS__) -#define svabs_s8_m(...) __builtin_sve_svabs_s8_m(__VA_ARGS__) -#define svabs_s32_m(...) __builtin_sve_svabs_s32_m(__VA_ARGS__) -#define svabs_s64_m(...) __builtin_sve_svabs_s64_m(__VA_ARGS__) -#define svabs_s16_m(...) __builtin_sve_svabs_s16_m(__VA_ARGS__) -#define svabs_s8_x(...) __builtin_sve_svabs_s8_x(__VA_ARGS__) -#define svabs_s32_x(...) __builtin_sve_svabs_s32_x(__VA_ARGS__) -#define svabs_s64_x(...) __builtin_sve_svabs_s64_x(__VA_ARGS__) -#define svabs_s16_x(...) __builtin_sve_svabs_s16_x(__VA_ARGS__) -#define svabs_s8_z(...) __builtin_sve_svabs_s8_z(__VA_ARGS__) -#define svabs_s32_z(...) __builtin_sve_svabs_s32_z(__VA_ARGS__) -#define svabs_s64_z(...) __builtin_sve_svabs_s64_z(__VA_ARGS__) -#define svabs_s16_z(...) __builtin_sve_svabs_s16_z(__VA_ARGS__) -#define svacge_n_f64(...) __builtin_sve_svacge_n_f64(__VA_ARGS__) -#define svacge_n_f32(...) __builtin_sve_svacge_n_f32(__VA_ARGS__) -#define svacge_n_f16(...) __builtin_sve_svacge_n_f16(__VA_ARGS__) -#define svacge_f64(...) __builtin_sve_svacge_f64(__VA_ARGS__) -#define svacge_f32(...) __builtin_sve_svacge_f32(__VA_ARGS__) -#define svacge_f16(...) __builtin_sve_svacge_f16(__VA_ARGS__) -#define svacgt_n_f64(...) __builtin_sve_svacgt_n_f64(__VA_ARGS__) -#define svacgt_n_f32(...) __builtin_sve_svacgt_n_f32(__VA_ARGS__) -#define svacgt_n_f16(...) __builtin_sve_svacgt_n_f16(__VA_ARGS__) -#define svacgt_f64(...) __builtin_sve_svacgt_f64(__VA_ARGS__) -#define svacgt_f32(...) __builtin_sve_svacgt_f32(__VA_ARGS__) -#define svacgt_f16(...) __builtin_sve_svacgt_f16(__VA_ARGS__) -#define svacle_n_f64(...) __builtin_sve_svacle_n_f64(__VA_ARGS__) -#define svacle_n_f32(...) __builtin_sve_svacle_n_f32(__VA_ARGS__) -#define svacle_n_f16(...) __builtin_sve_svacle_n_f16(__VA_ARGS__) -#define svacle_f64(...) __builtin_sve_svacle_f64(__VA_ARGS__) -#define svacle_f32(...) __builtin_sve_svacle_f32(__VA_ARGS__) -#define svacle_f16(...) __builtin_sve_svacle_f16(__VA_ARGS__) -#define svaclt_n_f64(...) __builtin_sve_svaclt_n_f64(__VA_ARGS__) -#define svaclt_n_f32(...) __builtin_sve_svaclt_n_f32(__VA_ARGS__) -#define svaclt_n_f16(...) __builtin_sve_svaclt_n_f16(__VA_ARGS__) -#define svaclt_f64(...) __builtin_sve_svaclt_f64(__VA_ARGS__) -#define svaclt_f32(...) __builtin_sve_svaclt_f32(__VA_ARGS__) -#define svaclt_f16(...) __builtin_sve_svaclt_f16(__VA_ARGS__) -#define svadd_n_f64_m(...) __builtin_sve_svadd_n_f64_m(__VA_ARGS__) -#define svadd_n_f32_m(...) __builtin_sve_svadd_n_f32_m(__VA_ARGS__) -#define svadd_n_f16_m(...) __builtin_sve_svadd_n_f16_m(__VA_ARGS__) -#define svadd_n_f64_x(...) __builtin_sve_svadd_n_f64_x(__VA_ARGS__) -#define svadd_n_f32_x(...) __builtin_sve_svadd_n_f32_x(__VA_ARGS__) -#define svadd_n_f16_x(...) __builtin_sve_svadd_n_f16_x(__VA_ARGS__) -#define svadd_n_f64_z(...) __builtin_sve_svadd_n_f64_z(__VA_ARGS__) -#define svadd_n_f32_z(...) __builtin_sve_svadd_n_f32_z(__VA_ARGS__) -#define svadd_n_f16_z(...) __builtin_sve_svadd_n_f16_z(__VA_ARGS__) -#define svadd_n_u8_m(...) __builtin_sve_svadd_n_u8_m(__VA_ARGS__) -#define svadd_n_u32_m(...) __builtin_sve_svadd_n_u32_m(__VA_ARGS__) -#define svadd_n_u64_m(...) __builtin_sve_svadd_n_u64_m(__VA_ARGS__) -#define svadd_n_u16_m(...) __builtin_sve_svadd_n_u16_m(__VA_ARGS__) -#define svadd_n_s8_m(...) __builtin_sve_svadd_n_s8_m(__VA_ARGS__) -#define svadd_n_s32_m(...) __builtin_sve_svadd_n_s32_m(__VA_ARGS__) -#define svadd_n_s64_m(...) __builtin_sve_svadd_n_s64_m(__VA_ARGS__) -#define svadd_n_s16_m(...) __builtin_sve_svadd_n_s16_m(__VA_ARGS__) -#define svadd_n_u8_x(...) __builtin_sve_svadd_n_u8_x(__VA_ARGS__) -#define svadd_n_u32_x(...) __builtin_sve_svadd_n_u32_x(__VA_ARGS__) -#define svadd_n_u64_x(...) __builtin_sve_svadd_n_u64_x(__VA_ARGS__) -#define svadd_n_u16_x(...) __builtin_sve_svadd_n_u16_x(__VA_ARGS__) -#define svadd_n_s8_x(...) __builtin_sve_svadd_n_s8_x(__VA_ARGS__) -#define svadd_n_s32_x(...) __builtin_sve_svadd_n_s32_x(__VA_ARGS__) -#define svadd_n_s64_x(...) __builtin_sve_svadd_n_s64_x(__VA_ARGS__) -#define svadd_n_s16_x(...) __builtin_sve_svadd_n_s16_x(__VA_ARGS__) -#define svadd_n_u8_z(...) __builtin_sve_svadd_n_u8_z(__VA_ARGS__) -#define svadd_n_u32_z(...) __builtin_sve_svadd_n_u32_z(__VA_ARGS__) -#define svadd_n_u64_z(...) __builtin_sve_svadd_n_u64_z(__VA_ARGS__) -#define svadd_n_u16_z(...) __builtin_sve_svadd_n_u16_z(__VA_ARGS__) -#define svadd_n_s8_z(...) __builtin_sve_svadd_n_s8_z(__VA_ARGS__) -#define svadd_n_s32_z(...) __builtin_sve_svadd_n_s32_z(__VA_ARGS__) -#define svadd_n_s64_z(...) __builtin_sve_svadd_n_s64_z(__VA_ARGS__) -#define svadd_n_s16_z(...) __builtin_sve_svadd_n_s16_z(__VA_ARGS__) -#define svadd_f64_m(...) __builtin_sve_svadd_f64_m(__VA_ARGS__) -#define svadd_f32_m(...) __builtin_sve_svadd_f32_m(__VA_ARGS__) -#define svadd_f16_m(...) __builtin_sve_svadd_f16_m(__VA_ARGS__) -#define svadd_f64_x(...) __builtin_sve_svadd_f64_x(__VA_ARGS__) -#define svadd_f32_x(...) __builtin_sve_svadd_f32_x(__VA_ARGS__) -#define svadd_f16_x(...) __builtin_sve_svadd_f16_x(__VA_ARGS__) -#define svadd_f64_z(...) __builtin_sve_svadd_f64_z(__VA_ARGS__) -#define svadd_f32_z(...) __builtin_sve_svadd_f32_z(__VA_ARGS__) -#define svadd_f16_z(...) __builtin_sve_svadd_f16_z(__VA_ARGS__) -#define svadd_u8_m(...) __builtin_sve_svadd_u8_m(__VA_ARGS__) -#define svadd_u32_m(...) __builtin_sve_svadd_u32_m(__VA_ARGS__) -#define svadd_u64_m(...) __builtin_sve_svadd_u64_m(__VA_ARGS__) -#define svadd_u16_m(...) __builtin_sve_svadd_u16_m(__VA_ARGS__) -#define svadd_s8_m(...) __builtin_sve_svadd_s8_m(__VA_ARGS__) -#define svadd_s32_m(...) __builtin_sve_svadd_s32_m(__VA_ARGS__) -#define svadd_s64_m(...) __builtin_sve_svadd_s64_m(__VA_ARGS__) -#define svadd_s16_m(...) __builtin_sve_svadd_s16_m(__VA_ARGS__) -#define svadd_u8_x(...) __builtin_sve_svadd_u8_x(__VA_ARGS__) -#define svadd_u32_x(...) __builtin_sve_svadd_u32_x(__VA_ARGS__) -#define svadd_u64_x(...) __builtin_sve_svadd_u64_x(__VA_ARGS__) -#define svadd_u16_x(...) __builtin_sve_svadd_u16_x(__VA_ARGS__) -#define svadd_s8_x(...) __builtin_sve_svadd_s8_x(__VA_ARGS__) -#define svadd_s32_x(...) __builtin_sve_svadd_s32_x(__VA_ARGS__) -#define svadd_s64_x(...) __builtin_sve_svadd_s64_x(__VA_ARGS__) -#define svadd_s16_x(...) __builtin_sve_svadd_s16_x(__VA_ARGS__) -#define svadd_u8_z(...) __builtin_sve_svadd_u8_z(__VA_ARGS__) -#define svadd_u32_z(...) __builtin_sve_svadd_u32_z(__VA_ARGS__) -#define svadd_u64_z(...) __builtin_sve_svadd_u64_z(__VA_ARGS__) -#define svadd_u16_z(...) __builtin_sve_svadd_u16_z(__VA_ARGS__) -#define svadd_s8_z(...) __builtin_sve_svadd_s8_z(__VA_ARGS__) -#define svadd_s32_z(...) __builtin_sve_svadd_s32_z(__VA_ARGS__) -#define svadd_s64_z(...) __builtin_sve_svadd_s64_z(__VA_ARGS__) -#define svadd_s16_z(...) __builtin_sve_svadd_s16_z(__VA_ARGS__) -#define svadda_f64(...) __builtin_sve_svadda_f64(__VA_ARGS__) -#define svadda_f32(...) __builtin_sve_svadda_f32(__VA_ARGS__) -#define svadda_f16(...) __builtin_sve_svadda_f16(__VA_ARGS__) -#define svaddv_s8(...) __builtin_sve_svaddv_s8(__VA_ARGS__) -#define svaddv_s32(...) __builtin_sve_svaddv_s32(__VA_ARGS__) -#define svaddv_s64(...) __builtin_sve_svaddv_s64(__VA_ARGS__) -#define svaddv_s16(...) __builtin_sve_svaddv_s16(__VA_ARGS__) -#define svaddv_u8(...) __builtin_sve_svaddv_u8(__VA_ARGS__) -#define svaddv_u32(...) __builtin_sve_svaddv_u32(__VA_ARGS__) -#define svaddv_u64(...) __builtin_sve_svaddv_u64(__VA_ARGS__) -#define svaddv_u16(...) __builtin_sve_svaddv_u16(__VA_ARGS__) -#define svaddv_f64(...) __builtin_sve_svaddv_f64(__VA_ARGS__) -#define svaddv_f32(...) __builtin_sve_svaddv_f32(__VA_ARGS__) -#define svaddv_f16(...) __builtin_sve_svaddv_f16(__VA_ARGS__) -#define svadrb_u32base_u32offset(...) __builtin_sve_svadrb_u32base_u32offset(__VA_ARGS__) -#define svadrb_u64base_u64offset(...) __builtin_sve_svadrb_u64base_u64offset(__VA_ARGS__) -#define svadrb_u32base_s32offset(...) __builtin_sve_svadrb_u32base_s32offset(__VA_ARGS__) -#define svadrb_u64base_s64offset(...) __builtin_sve_svadrb_u64base_s64offset(__VA_ARGS__) -#define svadrd_u32base_u32index(...) __builtin_sve_svadrd_u32base_u32index(__VA_ARGS__) -#define svadrd_u64base_u64index(...) __builtin_sve_svadrd_u64base_u64index(__VA_ARGS__) -#define svadrd_u32base_s32index(...) __builtin_sve_svadrd_u32base_s32index(__VA_ARGS__) -#define svadrd_u64base_s64index(...) __builtin_sve_svadrd_u64base_s64index(__VA_ARGS__) -#define svadrh_u32base_u32index(...) __builtin_sve_svadrh_u32base_u32index(__VA_ARGS__) -#define svadrh_u64base_u64index(...) __builtin_sve_svadrh_u64base_u64index(__VA_ARGS__) -#define svadrh_u32base_s32index(...) __builtin_sve_svadrh_u32base_s32index(__VA_ARGS__) -#define svadrh_u64base_s64index(...) __builtin_sve_svadrh_u64base_s64index(__VA_ARGS__) -#define svadrw_u32base_u32index(...) __builtin_sve_svadrw_u32base_u32index(__VA_ARGS__) -#define svadrw_u64base_u64index(...) __builtin_sve_svadrw_u64base_u64index(__VA_ARGS__) -#define svadrw_u32base_s32index(...) __builtin_sve_svadrw_u32base_s32index(__VA_ARGS__) -#define svadrw_u64base_s64index(...) __builtin_sve_svadrw_u64base_s64index(__VA_ARGS__) -#define svand_b_z(...) __builtin_sve_svand_b_z(__VA_ARGS__) -#define svand_n_u8_m(...) __builtin_sve_svand_n_u8_m(__VA_ARGS__) -#define svand_n_u32_m(...) __builtin_sve_svand_n_u32_m(__VA_ARGS__) -#define svand_n_u64_m(...) __builtin_sve_svand_n_u64_m(__VA_ARGS__) -#define svand_n_u16_m(...) __builtin_sve_svand_n_u16_m(__VA_ARGS__) -#define svand_n_s8_m(...) __builtin_sve_svand_n_s8_m(__VA_ARGS__) -#define svand_n_s32_m(...) __builtin_sve_svand_n_s32_m(__VA_ARGS__) -#define svand_n_s64_m(...) __builtin_sve_svand_n_s64_m(__VA_ARGS__) -#define svand_n_s16_m(...) __builtin_sve_svand_n_s16_m(__VA_ARGS__) -#define svand_n_u8_x(...) __builtin_sve_svand_n_u8_x(__VA_ARGS__) -#define svand_n_u32_x(...) __builtin_sve_svand_n_u32_x(__VA_ARGS__) -#define svand_n_u64_x(...) __builtin_sve_svand_n_u64_x(__VA_ARGS__) -#define svand_n_u16_x(...) __builtin_sve_svand_n_u16_x(__VA_ARGS__) -#define svand_n_s8_x(...) __builtin_sve_svand_n_s8_x(__VA_ARGS__) -#define svand_n_s32_x(...) __builtin_sve_svand_n_s32_x(__VA_ARGS__) -#define svand_n_s64_x(...) __builtin_sve_svand_n_s64_x(__VA_ARGS__) -#define svand_n_s16_x(...) __builtin_sve_svand_n_s16_x(__VA_ARGS__) -#define svand_n_u8_z(...) __builtin_sve_svand_n_u8_z(__VA_ARGS__) -#define svand_n_u32_z(...) __builtin_sve_svand_n_u32_z(__VA_ARGS__) -#define svand_n_u64_z(...) __builtin_sve_svand_n_u64_z(__VA_ARGS__) -#define svand_n_u16_z(...) __builtin_sve_svand_n_u16_z(__VA_ARGS__) -#define svand_n_s8_z(...) __builtin_sve_svand_n_s8_z(__VA_ARGS__) -#define svand_n_s32_z(...) __builtin_sve_svand_n_s32_z(__VA_ARGS__) -#define svand_n_s64_z(...) __builtin_sve_svand_n_s64_z(__VA_ARGS__) -#define svand_n_s16_z(...) __builtin_sve_svand_n_s16_z(__VA_ARGS__) -#define svand_u8_m(...) __builtin_sve_svand_u8_m(__VA_ARGS__) -#define svand_u32_m(...) __builtin_sve_svand_u32_m(__VA_ARGS__) -#define svand_u64_m(...) __builtin_sve_svand_u64_m(__VA_ARGS__) -#define svand_u16_m(...) __builtin_sve_svand_u16_m(__VA_ARGS__) -#define svand_s8_m(...) __builtin_sve_svand_s8_m(__VA_ARGS__) -#define svand_s32_m(...) __builtin_sve_svand_s32_m(__VA_ARGS__) -#define svand_s64_m(...) __builtin_sve_svand_s64_m(__VA_ARGS__) -#define svand_s16_m(...) __builtin_sve_svand_s16_m(__VA_ARGS__) -#define svand_u8_x(...) __builtin_sve_svand_u8_x(__VA_ARGS__) -#define svand_u32_x(...) __builtin_sve_svand_u32_x(__VA_ARGS__) -#define svand_u64_x(...) __builtin_sve_svand_u64_x(__VA_ARGS__) -#define svand_u16_x(...) __builtin_sve_svand_u16_x(__VA_ARGS__) -#define svand_s8_x(...) __builtin_sve_svand_s8_x(__VA_ARGS__) -#define svand_s32_x(...) __builtin_sve_svand_s32_x(__VA_ARGS__) -#define svand_s64_x(...) __builtin_sve_svand_s64_x(__VA_ARGS__) -#define svand_s16_x(...) __builtin_sve_svand_s16_x(__VA_ARGS__) -#define svand_u8_z(...) __builtin_sve_svand_u8_z(__VA_ARGS__) -#define svand_u32_z(...) __builtin_sve_svand_u32_z(__VA_ARGS__) -#define svand_u64_z(...) __builtin_sve_svand_u64_z(__VA_ARGS__) -#define svand_u16_z(...) __builtin_sve_svand_u16_z(__VA_ARGS__) -#define svand_s8_z(...) __builtin_sve_svand_s8_z(__VA_ARGS__) -#define svand_s32_z(...) __builtin_sve_svand_s32_z(__VA_ARGS__) -#define svand_s64_z(...) __builtin_sve_svand_s64_z(__VA_ARGS__) -#define svand_s16_z(...) __builtin_sve_svand_s16_z(__VA_ARGS__) -#define svandv_u8(...) __builtin_sve_svandv_u8(__VA_ARGS__) -#define svandv_u32(...) __builtin_sve_svandv_u32(__VA_ARGS__) -#define svandv_u64(...) __builtin_sve_svandv_u64(__VA_ARGS__) -#define svandv_u16(...) __builtin_sve_svandv_u16(__VA_ARGS__) -#define svandv_s8(...) __builtin_sve_svandv_s8(__VA_ARGS__) -#define svandv_s32(...) __builtin_sve_svandv_s32(__VA_ARGS__) -#define svandv_s64(...) __builtin_sve_svandv_s64(__VA_ARGS__) -#define svandv_s16(...) __builtin_sve_svandv_s16(__VA_ARGS__) -#define svasr_n_s8_m(...) __builtin_sve_svasr_n_s8_m(__VA_ARGS__) -#define svasr_n_s32_m(...) __builtin_sve_svasr_n_s32_m(__VA_ARGS__) -#define svasr_n_s64_m(...) __builtin_sve_svasr_n_s64_m(__VA_ARGS__) -#define svasr_n_s16_m(...) __builtin_sve_svasr_n_s16_m(__VA_ARGS__) -#define svasr_n_s8_x(...) __builtin_sve_svasr_n_s8_x(__VA_ARGS__) -#define svasr_n_s32_x(...) __builtin_sve_svasr_n_s32_x(__VA_ARGS__) -#define svasr_n_s64_x(...) __builtin_sve_svasr_n_s64_x(__VA_ARGS__) -#define svasr_n_s16_x(...) __builtin_sve_svasr_n_s16_x(__VA_ARGS__) -#define svasr_n_s8_z(...) __builtin_sve_svasr_n_s8_z(__VA_ARGS__) -#define svasr_n_s32_z(...) __builtin_sve_svasr_n_s32_z(__VA_ARGS__) -#define svasr_n_s64_z(...) __builtin_sve_svasr_n_s64_z(__VA_ARGS__) -#define svasr_n_s16_z(...) __builtin_sve_svasr_n_s16_z(__VA_ARGS__) -#define svasr_s8_m(...) __builtin_sve_svasr_s8_m(__VA_ARGS__) -#define svasr_s32_m(...) __builtin_sve_svasr_s32_m(__VA_ARGS__) -#define svasr_s64_m(...) __builtin_sve_svasr_s64_m(__VA_ARGS__) -#define svasr_s16_m(...) __builtin_sve_svasr_s16_m(__VA_ARGS__) -#define svasr_s8_x(...) __builtin_sve_svasr_s8_x(__VA_ARGS__) -#define svasr_s32_x(...) __builtin_sve_svasr_s32_x(__VA_ARGS__) -#define svasr_s64_x(...) __builtin_sve_svasr_s64_x(__VA_ARGS__) -#define svasr_s16_x(...) __builtin_sve_svasr_s16_x(__VA_ARGS__) -#define svasr_s8_z(...) __builtin_sve_svasr_s8_z(__VA_ARGS__) -#define svasr_s32_z(...) __builtin_sve_svasr_s32_z(__VA_ARGS__) -#define svasr_s64_z(...) __builtin_sve_svasr_s64_z(__VA_ARGS__) -#define svasr_s16_z(...) __builtin_sve_svasr_s16_z(__VA_ARGS__) -#define svasr_wide_n_s8_m(...) __builtin_sve_svasr_wide_n_s8_m(__VA_ARGS__) -#define svasr_wide_n_s32_m(...) __builtin_sve_svasr_wide_n_s32_m(__VA_ARGS__) -#define svasr_wide_n_s16_m(...) __builtin_sve_svasr_wide_n_s16_m(__VA_ARGS__) -#define svasr_wide_n_s8_x(...) __builtin_sve_svasr_wide_n_s8_x(__VA_ARGS__) -#define svasr_wide_n_s32_x(...) __builtin_sve_svasr_wide_n_s32_x(__VA_ARGS__) -#define svasr_wide_n_s16_x(...) __builtin_sve_svasr_wide_n_s16_x(__VA_ARGS__) -#define svasr_wide_n_s8_z(...) __builtin_sve_svasr_wide_n_s8_z(__VA_ARGS__) -#define svasr_wide_n_s32_z(...) __builtin_sve_svasr_wide_n_s32_z(__VA_ARGS__) -#define svasr_wide_n_s16_z(...) __builtin_sve_svasr_wide_n_s16_z(__VA_ARGS__) -#define svasr_wide_s8_m(...) __builtin_sve_svasr_wide_s8_m(__VA_ARGS__) -#define svasr_wide_s32_m(...) __builtin_sve_svasr_wide_s32_m(__VA_ARGS__) -#define svasr_wide_s16_m(...) __builtin_sve_svasr_wide_s16_m(__VA_ARGS__) -#define svasr_wide_s8_x(...) __builtin_sve_svasr_wide_s8_x(__VA_ARGS__) -#define svasr_wide_s32_x(...) __builtin_sve_svasr_wide_s32_x(__VA_ARGS__) -#define svasr_wide_s16_x(...) __builtin_sve_svasr_wide_s16_x(__VA_ARGS__) -#define svasr_wide_s8_z(...) __builtin_sve_svasr_wide_s8_z(__VA_ARGS__) -#define svasr_wide_s32_z(...) __builtin_sve_svasr_wide_s32_z(__VA_ARGS__) -#define svasr_wide_s16_z(...) __builtin_sve_svasr_wide_s16_z(__VA_ARGS__) -#define svasrd_n_s8_m(...) __builtin_sve_svasrd_n_s8_m(__VA_ARGS__) -#define svasrd_n_s32_m(...) __builtin_sve_svasrd_n_s32_m(__VA_ARGS__) -#define svasrd_n_s64_m(...) __builtin_sve_svasrd_n_s64_m(__VA_ARGS__) -#define svasrd_n_s16_m(...) __builtin_sve_svasrd_n_s16_m(__VA_ARGS__) -#define svasrd_n_s8_x(...) __builtin_sve_svasrd_n_s8_x(__VA_ARGS__) -#define svasrd_n_s32_x(...) __builtin_sve_svasrd_n_s32_x(__VA_ARGS__) -#define svasrd_n_s64_x(...) __builtin_sve_svasrd_n_s64_x(__VA_ARGS__) -#define svasrd_n_s16_x(...) __builtin_sve_svasrd_n_s16_x(__VA_ARGS__) -#define svasrd_n_s8_z(...) __builtin_sve_svasrd_n_s8_z(__VA_ARGS__) -#define svasrd_n_s32_z(...) __builtin_sve_svasrd_n_s32_z(__VA_ARGS__) -#define svasrd_n_s64_z(...) __builtin_sve_svasrd_n_s64_z(__VA_ARGS__) -#define svasrd_n_s16_z(...) __builtin_sve_svasrd_n_s16_z(__VA_ARGS__) -#define svbic_b_z(...) __builtin_sve_svbic_b_z(__VA_ARGS__) -#define svbic_n_u8_m(...) __builtin_sve_svbic_n_u8_m(__VA_ARGS__) -#define svbic_n_u32_m(...) __builtin_sve_svbic_n_u32_m(__VA_ARGS__) -#define svbic_n_u64_m(...) __builtin_sve_svbic_n_u64_m(__VA_ARGS__) -#define svbic_n_u16_m(...) __builtin_sve_svbic_n_u16_m(__VA_ARGS__) -#define svbic_n_s8_m(...) __builtin_sve_svbic_n_s8_m(__VA_ARGS__) -#define svbic_n_s32_m(...) __builtin_sve_svbic_n_s32_m(__VA_ARGS__) -#define svbic_n_s64_m(...) __builtin_sve_svbic_n_s64_m(__VA_ARGS__) -#define svbic_n_s16_m(...) __builtin_sve_svbic_n_s16_m(__VA_ARGS__) -#define svbic_n_u8_x(...) __builtin_sve_svbic_n_u8_x(__VA_ARGS__) -#define svbic_n_u32_x(...) __builtin_sve_svbic_n_u32_x(__VA_ARGS__) -#define svbic_n_u64_x(...) __builtin_sve_svbic_n_u64_x(__VA_ARGS__) -#define svbic_n_u16_x(...) __builtin_sve_svbic_n_u16_x(__VA_ARGS__) -#define svbic_n_s8_x(...) __builtin_sve_svbic_n_s8_x(__VA_ARGS__) -#define svbic_n_s32_x(...) __builtin_sve_svbic_n_s32_x(__VA_ARGS__) -#define svbic_n_s64_x(...) __builtin_sve_svbic_n_s64_x(__VA_ARGS__) -#define svbic_n_s16_x(...) __builtin_sve_svbic_n_s16_x(__VA_ARGS__) -#define svbic_n_u8_z(...) __builtin_sve_svbic_n_u8_z(__VA_ARGS__) -#define svbic_n_u32_z(...) __builtin_sve_svbic_n_u32_z(__VA_ARGS__) -#define svbic_n_u64_z(...) __builtin_sve_svbic_n_u64_z(__VA_ARGS__) -#define svbic_n_u16_z(...) __builtin_sve_svbic_n_u16_z(__VA_ARGS__) -#define svbic_n_s8_z(...) __builtin_sve_svbic_n_s8_z(__VA_ARGS__) -#define svbic_n_s32_z(...) __builtin_sve_svbic_n_s32_z(__VA_ARGS__) -#define svbic_n_s64_z(...) __builtin_sve_svbic_n_s64_z(__VA_ARGS__) -#define svbic_n_s16_z(...) __builtin_sve_svbic_n_s16_z(__VA_ARGS__) -#define svbic_u8_m(...) __builtin_sve_svbic_u8_m(__VA_ARGS__) -#define svbic_u32_m(...) __builtin_sve_svbic_u32_m(__VA_ARGS__) -#define svbic_u64_m(...) __builtin_sve_svbic_u64_m(__VA_ARGS__) -#define svbic_u16_m(...) __builtin_sve_svbic_u16_m(__VA_ARGS__) -#define svbic_s8_m(...) __builtin_sve_svbic_s8_m(__VA_ARGS__) -#define svbic_s32_m(...) __builtin_sve_svbic_s32_m(__VA_ARGS__) -#define svbic_s64_m(...) __builtin_sve_svbic_s64_m(__VA_ARGS__) -#define svbic_s16_m(...) __builtin_sve_svbic_s16_m(__VA_ARGS__) -#define svbic_u8_x(...) __builtin_sve_svbic_u8_x(__VA_ARGS__) -#define svbic_u32_x(...) __builtin_sve_svbic_u32_x(__VA_ARGS__) -#define svbic_u64_x(...) __builtin_sve_svbic_u64_x(__VA_ARGS__) -#define svbic_u16_x(...) __builtin_sve_svbic_u16_x(__VA_ARGS__) -#define svbic_s8_x(...) __builtin_sve_svbic_s8_x(__VA_ARGS__) -#define svbic_s32_x(...) __builtin_sve_svbic_s32_x(__VA_ARGS__) -#define svbic_s64_x(...) __builtin_sve_svbic_s64_x(__VA_ARGS__) -#define svbic_s16_x(...) __builtin_sve_svbic_s16_x(__VA_ARGS__) -#define svbic_u8_z(...) __builtin_sve_svbic_u8_z(__VA_ARGS__) -#define svbic_u32_z(...) __builtin_sve_svbic_u32_z(__VA_ARGS__) -#define svbic_u64_z(...) __builtin_sve_svbic_u64_z(__VA_ARGS__) -#define svbic_u16_z(...) __builtin_sve_svbic_u16_z(__VA_ARGS__) -#define svbic_s8_z(...) __builtin_sve_svbic_s8_z(__VA_ARGS__) -#define svbic_s32_z(...) __builtin_sve_svbic_s32_z(__VA_ARGS__) -#define svbic_s64_z(...) __builtin_sve_svbic_s64_z(__VA_ARGS__) -#define svbic_s16_z(...) __builtin_sve_svbic_s16_z(__VA_ARGS__) -#define svbrka_b_m(...) __builtin_sve_svbrka_b_m(__VA_ARGS__) -#define svbrka_b_z(...) __builtin_sve_svbrka_b_z(__VA_ARGS__) -#define svbrkb_b_m(...) __builtin_sve_svbrkb_b_m(__VA_ARGS__) -#define svbrkb_b_z(...) __builtin_sve_svbrkb_b_z(__VA_ARGS__) -#define svbrkn_b_z(...) __builtin_sve_svbrkn_b_z(__VA_ARGS__) -#define svbrkpa_b_z(...) __builtin_sve_svbrkpa_b_z(__VA_ARGS__) -#define svbrkpb_b_z(...) __builtin_sve_svbrkpb_b_z(__VA_ARGS__) -#define svcadd_f64_m(...) __builtin_sve_svcadd_f64_m(__VA_ARGS__) -#define svcadd_f32_m(...) __builtin_sve_svcadd_f32_m(__VA_ARGS__) -#define svcadd_f16_m(...) __builtin_sve_svcadd_f16_m(__VA_ARGS__) -#define svcadd_f64_x(...) __builtin_sve_svcadd_f64_x(__VA_ARGS__) -#define svcadd_f32_x(...) __builtin_sve_svcadd_f32_x(__VA_ARGS__) -#define svcadd_f16_x(...) __builtin_sve_svcadd_f16_x(__VA_ARGS__) -#define svcadd_f64_z(...) __builtin_sve_svcadd_f64_z(__VA_ARGS__) -#define svcadd_f32_z(...) __builtin_sve_svcadd_f32_z(__VA_ARGS__) -#define svcadd_f16_z(...) __builtin_sve_svcadd_f16_z(__VA_ARGS__) -#define svclasta_n_u8(...) __builtin_sve_svclasta_n_u8(__VA_ARGS__) -#define svclasta_n_u32(...) __builtin_sve_svclasta_n_u32(__VA_ARGS__) -#define svclasta_n_u64(...) __builtin_sve_svclasta_n_u64(__VA_ARGS__) -#define svclasta_n_u16(...) __builtin_sve_svclasta_n_u16(__VA_ARGS__) -#define svclasta_n_s8(...) __builtin_sve_svclasta_n_s8(__VA_ARGS__) -#define svclasta_n_f64(...) __builtin_sve_svclasta_n_f64(__VA_ARGS__) -#define svclasta_n_f32(...) __builtin_sve_svclasta_n_f32(__VA_ARGS__) -#define svclasta_n_f16(...) __builtin_sve_svclasta_n_f16(__VA_ARGS__) -#define svclasta_n_s32(...) __builtin_sve_svclasta_n_s32(__VA_ARGS__) -#define svclasta_n_s64(...) __builtin_sve_svclasta_n_s64(__VA_ARGS__) -#define svclasta_n_s16(...) __builtin_sve_svclasta_n_s16(__VA_ARGS__) -#define svclasta_u8(...) __builtin_sve_svclasta_u8(__VA_ARGS__) -#define svclasta_u32(...) __builtin_sve_svclasta_u32(__VA_ARGS__) -#define svclasta_u64(...) __builtin_sve_svclasta_u64(__VA_ARGS__) -#define svclasta_u16(...) __builtin_sve_svclasta_u16(__VA_ARGS__) -#define svclasta_s8(...) __builtin_sve_svclasta_s8(__VA_ARGS__) -#define svclasta_f64(...) __builtin_sve_svclasta_f64(__VA_ARGS__) -#define svclasta_f32(...) __builtin_sve_svclasta_f32(__VA_ARGS__) -#define svclasta_f16(...) __builtin_sve_svclasta_f16(__VA_ARGS__) -#define svclasta_s32(...) __builtin_sve_svclasta_s32(__VA_ARGS__) -#define svclasta_s64(...) __builtin_sve_svclasta_s64(__VA_ARGS__) -#define svclasta_s16(...) __builtin_sve_svclasta_s16(__VA_ARGS__) -#define svclastb_n_u8(...) __builtin_sve_svclastb_n_u8(__VA_ARGS__) -#define svclastb_n_u32(...) __builtin_sve_svclastb_n_u32(__VA_ARGS__) -#define svclastb_n_u64(...) __builtin_sve_svclastb_n_u64(__VA_ARGS__) -#define svclastb_n_u16(...) __builtin_sve_svclastb_n_u16(__VA_ARGS__) -#define svclastb_n_s8(...) __builtin_sve_svclastb_n_s8(__VA_ARGS__) -#define svclastb_n_f64(...) __builtin_sve_svclastb_n_f64(__VA_ARGS__) -#define svclastb_n_f32(...) __builtin_sve_svclastb_n_f32(__VA_ARGS__) -#define svclastb_n_f16(...) __builtin_sve_svclastb_n_f16(__VA_ARGS__) -#define svclastb_n_s32(...) __builtin_sve_svclastb_n_s32(__VA_ARGS__) -#define svclastb_n_s64(...) __builtin_sve_svclastb_n_s64(__VA_ARGS__) -#define svclastb_n_s16(...) __builtin_sve_svclastb_n_s16(__VA_ARGS__) -#define svclastb_u8(...) __builtin_sve_svclastb_u8(__VA_ARGS__) -#define svclastb_u32(...) __builtin_sve_svclastb_u32(__VA_ARGS__) -#define svclastb_u64(...) __builtin_sve_svclastb_u64(__VA_ARGS__) -#define svclastb_u16(...) __builtin_sve_svclastb_u16(__VA_ARGS__) -#define svclastb_s8(...) __builtin_sve_svclastb_s8(__VA_ARGS__) -#define svclastb_f64(...) __builtin_sve_svclastb_f64(__VA_ARGS__) -#define svclastb_f32(...) __builtin_sve_svclastb_f32(__VA_ARGS__) -#define svclastb_f16(...) __builtin_sve_svclastb_f16(__VA_ARGS__) -#define svclastb_s32(...) __builtin_sve_svclastb_s32(__VA_ARGS__) -#define svclastb_s64(...) __builtin_sve_svclastb_s64(__VA_ARGS__) -#define svclastb_s16(...) __builtin_sve_svclastb_s16(__VA_ARGS__) -#define svcls_s8_m(...) __builtin_sve_svcls_s8_m(__VA_ARGS__) -#define svcls_s32_m(...) __builtin_sve_svcls_s32_m(__VA_ARGS__) -#define svcls_s64_m(...) __builtin_sve_svcls_s64_m(__VA_ARGS__) -#define svcls_s16_m(...) __builtin_sve_svcls_s16_m(__VA_ARGS__) -#define svcls_s8_x(...) __builtin_sve_svcls_s8_x(__VA_ARGS__) -#define svcls_s32_x(...) __builtin_sve_svcls_s32_x(__VA_ARGS__) -#define svcls_s64_x(...) __builtin_sve_svcls_s64_x(__VA_ARGS__) -#define svcls_s16_x(...) __builtin_sve_svcls_s16_x(__VA_ARGS__) -#define svcls_s8_z(...) __builtin_sve_svcls_s8_z(__VA_ARGS__) -#define svcls_s32_z(...) __builtin_sve_svcls_s32_z(__VA_ARGS__) -#define svcls_s64_z(...) __builtin_sve_svcls_s64_z(__VA_ARGS__) -#define svcls_s16_z(...) __builtin_sve_svcls_s16_z(__VA_ARGS__) -#define svclz_u8_m(...) __builtin_sve_svclz_u8_m(__VA_ARGS__) -#define svclz_u32_m(...) __builtin_sve_svclz_u32_m(__VA_ARGS__) -#define svclz_u64_m(...) __builtin_sve_svclz_u64_m(__VA_ARGS__) -#define svclz_u16_m(...) __builtin_sve_svclz_u16_m(__VA_ARGS__) -#define svclz_s8_m(...) __builtin_sve_svclz_s8_m(__VA_ARGS__) -#define svclz_s32_m(...) __builtin_sve_svclz_s32_m(__VA_ARGS__) -#define svclz_s64_m(...) __builtin_sve_svclz_s64_m(__VA_ARGS__) -#define svclz_s16_m(...) __builtin_sve_svclz_s16_m(__VA_ARGS__) -#define svclz_u8_x(...) __builtin_sve_svclz_u8_x(__VA_ARGS__) -#define svclz_u32_x(...) __builtin_sve_svclz_u32_x(__VA_ARGS__) -#define svclz_u64_x(...) __builtin_sve_svclz_u64_x(__VA_ARGS__) -#define svclz_u16_x(...) __builtin_sve_svclz_u16_x(__VA_ARGS__) -#define svclz_s8_x(...) __builtin_sve_svclz_s8_x(__VA_ARGS__) -#define svclz_s32_x(...) __builtin_sve_svclz_s32_x(__VA_ARGS__) -#define svclz_s64_x(...) __builtin_sve_svclz_s64_x(__VA_ARGS__) -#define svclz_s16_x(...) __builtin_sve_svclz_s16_x(__VA_ARGS__) -#define svclz_u8_z(...) __builtin_sve_svclz_u8_z(__VA_ARGS__) -#define svclz_u32_z(...) __builtin_sve_svclz_u32_z(__VA_ARGS__) -#define svclz_u64_z(...) __builtin_sve_svclz_u64_z(__VA_ARGS__) -#define svclz_u16_z(...) __builtin_sve_svclz_u16_z(__VA_ARGS__) -#define svclz_s8_z(...) __builtin_sve_svclz_s8_z(__VA_ARGS__) -#define svclz_s32_z(...) __builtin_sve_svclz_s32_z(__VA_ARGS__) -#define svclz_s64_z(...) __builtin_sve_svclz_s64_z(__VA_ARGS__) -#define svclz_s16_z(...) __builtin_sve_svclz_s16_z(__VA_ARGS__) -#define svcmla_f64_m(...) __builtin_sve_svcmla_f64_m(__VA_ARGS__) -#define svcmla_f32_m(...) __builtin_sve_svcmla_f32_m(__VA_ARGS__) -#define svcmla_f16_m(...) __builtin_sve_svcmla_f16_m(__VA_ARGS__) -#define svcmla_f64_x(...) __builtin_sve_svcmla_f64_x(__VA_ARGS__) -#define svcmla_f32_x(...) __builtin_sve_svcmla_f32_x(__VA_ARGS__) -#define svcmla_f16_x(...) __builtin_sve_svcmla_f16_x(__VA_ARGS__) -#define svcmla_f64_z(...) __builtin_sve_svcmla_f64_z(__VA_ARGS__) -#define svcmla_f32_z(...) __builtin_sve_svcmla_f32_z(__VA_ARGS__) -#define svcmla_f16_z(...) __builtin_sve_svcmla_f16_z(__VA_ARGS__) -#define svcmla_lane_f32(...) __builtin_sve_svcmla_lane_f32(__VA_ARGS__) -#define svcmla_lane_f16(...) __builtin_sve_svcmla_lane_f16(__VA_ARGS__) -#define svcmpeq_n_f64(...) __builtin_sve_svcmpeq_n_f64(__VA_ARGS__) -#define svcmpeq_n_f32(...) __builtin_sve_svcmpeq_n_f32(__VA_ARGS__) -#define svcmpeq_n_f16(...) __builtin_sve_svcmpeq_n_f16(__VA_ARGS__) -#define svcmpeq_n_u8(...) __builtin_sve_svcmpeq_n_u8(__VA_ARGS__) -#define svcmpeq_n_u32(...) __builtin_sve_svcmpeq_n_u32(__VA_ARGS__) -#define svcmpeq_n_u64(...) __builtin_sve_svcmpeq_n_u64(__VA_ARGS__) -#define svcmpeq_n_u16(...) __builtin_sve_svcmpeq_n_u16(__VA_ARGS__) -#define svcmpeq_n_s8(...) __builtin_sve_svcmpeq_n_s8(__VA_ARGS__) -#define svcmpeq_n_s32(...) __builtin_sve_svcmpeq_n_s32(__VA_ARGS__) -#define svcmpeq_n_s64(...) __builtin_sve_svcmpeq_n_s64(__VA_ARGS__) -#define svcmpeq_n_s16(...) __builtin_sve_svcmpeq_n_s16(__VA_ARGS__) -#define svcmpeq_u8(...) __builtin_sve_svcmpeq_u8(__VA_ARGS__) -#define svcmpeq_u32(...) __builtin_sve_svcmpeq_u32(__VA_ARGS__) -#define svcmpeq_u64(...) __builtin_sve_svcmpeq_u64(__VA_ARGS__) -#define svcmpeq_u16(...) __builtin_sve_svcmpeq_u16(__VA_ARGS__) -#define svcmpeq_s8(...) __builtin_sve_svcmpeq_s8(__VA_ARGS__) -#define svcmpeq_s32(...) __builtin_sve_svcmpeq_s32(__VA_ARGS__) -#define svcmpeq_s64(...) __builtin_sve_svcmpeq_s64(__VA_ARGS__) -#define svcmpeq_s16(...) __builtin_sve_svcmpeq_s16(__VA_ARGS__) -#define svcmpeq_f64(...) __builtin_sve_svcmpeq_f64(__VA_ARGS__) -#define svcmpeq_f32(...) __builtin_sve_svcmpeq_f32(__VA_ARGS__) -#define svcmpeq_f16(...) __builtin_sve_svcmpeq_f16(__VA_ARGS__) -#define svcmpeq_wide_n_s8(...) __builtin_sve_svcmpeq_wide_n_s8(__VA_ARGS__) -#define svcmpeq_wide_n_s32(...) __builtin_sve_svcmpeq_wide_n_s32(__VA_ARGS__) -#define svcmpeq_wide_n_s16(...) __builtin_sve_svcmpeq_wide_n_s16(__VA_ARGS__) -#define svcmpeq_wide_s8(...) __builtin_sve_svcmpeq_wide_s8(__VA_ARGS__) -#define svcmpeq_wide_s32(...) __builtin_sve_svcmpeq_wide_s32(__VA_ARGS__) -#define svcmpeq_wide_s16(...) __builtin_sve_svcmpeq_wide_s16(__VA_ARGS__) -#define svcmpge_n_f64(...) __builtin_sve_svcmpge_n_f64(__VA_ARGS__) -#define svcmpge_n_f32(...) __builtin_sve_svcmpge_n_f32(__VA_ARGS__) -#define svcmpge_n_f16(...) __builtin_sve_svcmpge_n_f16(__VA_ARGS__) -#define svcmpge_n_s8(...) __builtin_sve_svcmpge_n_s8(__VA_ARGS__) -#define svcmpge_n_s32(...) __builtin_sve_svcmpge_n_s32(__VA_ARGS__) -#define svcmpge_n_s64(...) __builtin_sve_svcmpge_n_s64(__VA_ARGS__) -#define svcmpge_n_s16(...) __builtin_sve_svcmpge_n_s16(__VA_ARGS__) -#define svcmpge_n_u8(...) __builtin_sve_svcmpge_n_u8(__VA_ARGS__) -#define svcmpge_n_u32(...) __builtin_sve_svcmpge_n_u32(__VA_ARGS__) -#define svcmpge_n_u64(...) __builtin_sve_svcmpge_n_u64(__VA_ARGS__) -#define svcmpge_n_u16(...) __builtin_sve_svcmpge_n_u16(__VA_ARGS__) -#define svcmpge_s8(...) __builtin_sve_svcmpge_s8(__VA_ARGS__) -#define svcmpge_s32(...) __builtin_sve_svcmpge_s32(__VA_ARGS__) -#define svcmpge_s64(...) __builtin_sve_svcmpge_s64(__VA_ARGS__) -#define svcmpge_s16(...) __builtin_sve_svcmpge_s16(__VA_ARGS__) -#define svcmpge_f64(...) __builtin_sve_svcmpge_f64(__VA_ARGS__) -#define svcmpge_f32(...) __builtin_sve_svcmpge_f32(__VA_ARGS__) -#define svcmpge_f16(...) __builtin_sve_svcmpge_f16(__VA_ARGS__) -#define svcmpge_u8(...) __builtin_sve_svcmpge_u8(__VA_ARGS__) -#define svcmpge_u32(...) __builtin_sve_svcmpge_u32(__VA_ARGS__) -#define svcmpge_u64(...) __builtin_sve_svcmpge_u64(__VA_ARGS__) -#define svcmpge_u16(...) __builtin_sve_svcmpge_u16(__VA_ARGS__) -#define svcmpge_wide_n_s8(...) __builtin_sve_svcmpge_wide_n_s8(__VA_ARGS__) -#define svcmpge_wide_n_s32(...) __builtin_sve_svcmpge_wide_n_s32(__VA_ARGS__) -#define svcmpge_wide_n_s16(...) __builtin_sve_svcmpge_wide_n_s16(__VA_ARGS__) -#define svcmpge_wide_n_u8(...) __builtin_sve_svcmpge_wide_n_u8(__VA_ARGS__) -#define svcmpge_wide_n_u32(...) __builtin_sve_svcmpge_wide_n_u32(__VA_ARGS__) -#define svcmpge_wide_n_u16(...) __builtin_sve_svcmpge_wide_n_u16(__VA_ARGS__) -#define svcmpge_wide_s8(...) __builtin_sve_svcmpge_wide_s8(__VA_ARGS__) -#define svcmpge_wide_s32(...) __builtin_sve_svcmpge_wide_s32(__VA_ARGS__) -#define svcmpge_wide_s16(...) __builtin_sve_svcmpge_wide_s16(__VA_ARGS__) -#define svcmpge_wide_u8(...) __builtin_sve_svcmpge_wide_u8(__VA_ARGS__) -#define svcmpge_wide_u32(...) __builtin_sve_svcmpge_wide_u32(__VA_ARGS__) -#define svcmpge_wide_u16(...) __builtin_sve_svcmpge_wide_u16(__VA_ARGS__) -#define svcmpgt_n_f64(...) __builtin_sve_svcmpgt_n_f64(__VA_ARGS__) -#define svcmpgt_n_f32(...) __builtin_sve_svcmpgt_n_f32(__VA_ARGS__) -#define svcmpgt_n_f16(...) __builtin_sve_svcmpgt_n_f16(__VA_ARGS__) -#define svcmpgt_n_s8(...) __builtin_sve_svcmpgt_n_s8(__VA_ARGS__) -#define svcmpgt_n_s32(...) __builtin_sve_svcmpgt_n_s32(__VA_ARGS__) -#define svcmpgt_n_s64(...) __builtin_sve_svcmpgt_n_s64(__VA_ARGS__) -#define svcmpgt_n_s16(...) __builtin_sve_svcmpgt_n_s16(__VA_ARGS__) -#define svcmpgt_n_u8(...) __builtin_sve_svcmpgt_n_u8(__VA_ARGS__) -#define svcmpgt_n_u32(...) __builtin_sve_svcmpgt_n_u32(__VA_ARGS__) -#define svcmpgt_n_u64(...) __builtin_sve_svcmpgt_n_u64(__VA_ARGS__) -#define svcmpgt_n_u16(...) __builtin_sve_svcmpgt_n_u16(__VA_ARGS__) -#define svcmpgt_s8(...) __builtin_sve_svcmpgt_s8(__VA_ARGS__) -#define svcmpgt_s32(...) __builtin_sve_svcmpgt_s32(__VA_ARGS__) -#define svcmpgt_s64(...) __builtin_sve_svcmpgt_s64(__VA_ARGS__) -#define svcmpgt_s16(...) __builtin_sve_svcmpgt_s16(__VA_ARGS__) -#define svcmpgt_f64(...) __builtin_sve_svcmpgt_f64(__VA_ARGS__) -#define svcmpgt_f32(...) __builtin_sve_svcmpgt_f32(__VA_ARGS__) -#define svcmpgt_f16(...) __builtin_sve_svcmpgt_f16(__VA_ARGS__) -#define svcmpgt_u8(...) __builtin_sve_svcmpgt_u8(__VA_ARGS__) -#define svcmpgt_u32(...) __builtin_sve_svcmpgt_u32(__VA_ARGS__) -#define svcmpgt_u64(...) __builtin_sve_svcmpgt_u64(__VA_ARGS__) -#define svcmpgt_u16(...) __builtin_sve_svcmpgt_u16(__VA_ARGS__) -#define svcmpgt_wide_n_s8(...) __builtin_sve_svcmpgt_wide_n_s8(__VA_ARGS__) -#define svcmpgt_wide_n_s32(...) __builtin_sve_svcmpgt_wide_n_s32(__VA_ARGS__) -#define svcmpgt_wide_n_s16(...) __builtin_sve_svcmpgt_wide_n_s16(__VA_ARGS__) -#define svcmpgt_wide_n_u8(...) __builtin_sve_svcmpgt_wide_n_u8(__VA_ARGS__) -#define svcmpgt_wide_n_u32(...) __builtin_sve_svcmpgt_wide_n_u32(__VA_ARGS__) -#define svcmpgt_wide_n_u16(...) __builtin_sve_svcmpgt_wide_n_u16(__VA_ARGS__) -#define svcmpgt_wide_s8(...) __builtin_sve_svcmpgt_wide_s8(__VA_ARGS__) -#define svcmpgt_wide_s32(...) __builtin_sve_svcmpgt_wide_s32(__VA_ARGS__) -#define svcmpgt_wide_s16(...) __builtin_sve_svcmpgt_wide_s16(__VA_ARGS__) -#define svcmpgt_wide_u8(...) __builtin_sve_svcmpgt_wide_u8(__VA_ARGS__) -#define svcmpgt_wide_u32(...) __builtin_sve_svcmpgt_wide_u32(__VA_ARGS__) -#define svcmpgt_wide_u16(...) __builtin_sve_svcmpgt_wide_u16(__VA_ARGS__) -#define svcmple_n_f64(...) __builtin_sve_svcmple_n_f64(__VA_ARGS__) -#define svcmple_n_f32(...) __builtin_sve_svcmple_n_f32(__VA_ARGS__) -#define svcmple_n_f16(...) __builtin_sve_svcmple_n_f16(__VA_ARGS__) -#define svcmple_n_s8(...) __builtin_sve_svcmple_n_s8(__VA_ARGS__) -#define svcmple_n_s32(...) __builtin_sve_svcmple_n_s32(__VA_ARGS__) -#define svcmple_n_s64(...) __builtin_sve_svcmple_n_s64(__VA_ARGS__) -#define svcmple_n_s16(...) __builtin_sve_svcmple_n_s16(__VA_ARGS__) -#define svcmple_n_u8(...) __builtin_sve_svcmple_n_u8(__VA_ARGS__) -#define svcmple_n_u32(...) __builtin_sve_svcmple_n_u32(__VA_ARGS__) -#define svcmple_n_u64(...) __builtin_sve_svcmple_n_u64(__VA_ARGS__) -#define svcmple_n_u16(...) __builtin_sve_svcmple_n_u16(__VA_ARGS__) -#define svcmple_s8(...) __builtin_sve_svcmple_s8(__VA_ARGS__) -#define svcmple_s32(...) __builtin_sve_svcmple_s32(__VA_ARGS__) -#define svcmple_s64(...) __builtin_sve_svcmple_s64(__VA_ARGS__) -#define svcmple_s16(...) __builtin_sve_svcmple_s16(__VA_ARGS__) -#define svcmple_f64(...) __builtin_sve_svcmple_f64(__VA_ARGS__) -#define svcmple_f32(...) __builtin_sve_svcmple_f32(__VA_ARGS__) -#define svcmple_f16(...) __builtin_sve_svcmple_f16(__VA_ARGS__) -#define svcmple_u8(...) __builtin_sve_svcmple_u8(__VA_ARGS__) -#define svcmple_u32(...) __builtin_sve_svcmple_u32(__VA_ARGS__) -#define svcmple_u64(...) __builtin_sve_svcmple_u64(__VA_ARGS__) -#define svcmple_u16(...) __builtin_sve_svcmple_u16(__VA_ARGS__) -#define svcmple_wide_n_s8(...) __builtin_sve_svcmple_wide_n_s8(__VA_ARGS__) -#define svcmple_wide_n_s32(...) __builtin_sve_svcmple_wide_n_s32(__VA_ARGS__) -#define svcmple_wide_n_s16(...) __builtin_sve_svcmple_wide_n_s16(__VA_ARGS__) -#define svcmple_wide_n_u8(...) __builtin_sve_svcmple_wide_n_u8(__VA_ARGS__) -#define svcmple_wide_n_u32(...) __builtin_sve_svcmple_wide_n_u32(__VA_ARGS__) -#define svcmple_wide_n_u16(...) __builtin_sve_svcmple_wide_n_u16(__VA_ARGS__) -#define svcmple_wide_s8(...) __builtin_sve_svcmple_wide_s8(__VA_ARGS__) -#define svcmple_wide_s32(...) __builtin_sve_svcmple_wide_s32(__VA_ARGS__) -#define svcmple_wide_s16(...) __builtin_sve_svcmple_wide_s16(__VA_ARGS__) -#define svcmple_wide_u8(...) __builtin_sve_svcmple_wide_u8(__VA_ARGS__) -#define svcmple_wide_u32(...) __builtin_sve_svcmple_wide_u32(__VA_ARGS__) -#define svcmple_wide_u16(...) __builtin_sve_svcmple_wide_u16(__VA_ARGS__) -#define svcmplt_n_u8(...) __builtin_sve_svcmplt_n_u8(__VA_ARGS__) -#define svcmplt_n_u32(...) __builtin_sve_svcmplt_n_u32(__VA_ARGS__) -#define svcmplt_n_u64(...) __builtin_sve_svcmplt_n_u64(__VA_ARGS__) -#define svcmplt_n_u16(...) __builtin_sve_svcmplt_n_u16(__VA_ARGS__) -#define svcmplt_n_f64(...) __builtin_sve_svcmplt_n_f64(__VA_ARGS__) -#define svcmplt_n_f32(...) __builtin_sve_svcmplt_n_f32(__VA_ARGS__) -#define svcmplt_n_f16(...) __builtin_sve_svcmplt_n_f16(__VA_ARGS__) -#define svcmplt_n_s8(...) __builtin_sve_svcmplt_n_s8(__VA_ARGS__) -#define svcmplt_n_s32(...) __builtin_sve_svcmplt_n_s32(__VA_ARGS__) -#define svcmplt_n_s64(...) __builtin_sve_svcmplt_n_s64(__VA_ARGS__) -#define svcmplt_n_s16(...) __builtin_sve_svcmplt_n_s16(__VA_ARGS__) -#define svcmplt_u8(...) __builtin_sve_svcmplt_u8(__VA_ARGS__) -#define svcmplt_u32(...) __builtin_sve_svcmplt_u32(__VA_ARGS__) -#define svcmplt_u64(...) __builtin_sve_svcmplt_u64(__VA_ARGS__) -#define svcmplt_u16(...) __builtin_sve_svcmplt_u16(__VA_ARGS__) -#define svcmplt_s8(...) __builtin_sve_svcmplt_s8(__VA_ARGS__) -#define svcmplt_s32(...) __builtin_sve_svcmplt_s32(__VA_ARGS__) -#define svcmplt_s64(...) __builtin_sve_svcmplt_s64(__VA_ARGS__) -#define svcmplt_s16(...) __builtin_sve_svcmplt_s16(__VA_ARGS__) -#define svcmplt_f64(...) __builtin_sve_svcmplt_f64(__VA_ARGS__) -#define svcmplt_f32(...) __builtin_sve_svcmplt_f32(__VA_ARGS__) -#define svcmplt_f16(...) __builtin_sve_svcmplt_f16(__VA_ARGS__) -#define svcmplt_wide_n_u8(...) __builtin_sve_svcmplt_wide_n_u8(__VA_ARGS__) -#define svcmplt_wide_n_u32(...) __builtin_sve_svcmplt_wide_n_u32(__VA_ARGS__) -#define svcmplt_wide_n_u16(...) __builtin_sve_svcmplt_wide_n_u16(__VA_ARGS__) -#define svcmplt_wide_n_s8(...) __builtin_sve_svcmplt_wide_n_s8(__VA_ARGS__) -#define svcmplt_wide_n_s32(...) __builtin_sve_svcmplt_wide_n_s32(__VA_ARGS__) -#define svcmplt_wide_n_s16(...) __builtin_sve_svcmplt_wide_n_s16(__VA_ARGS__) -#define svcmplt_wide_u8(...) __builtin_sve_svcmplt_wide_u8(__VA_ARGS__) -#define svcmplt_wide_u32(...) __builtin_sve_svcmplt_wide_u32(__VA_ARGS__) -#define svcmplt_wide_u16(...) __builtin_sve_svcmplt_wide_u16(__VA_ARGS__) -#define svcmplt_wide_s8(...) __builtin_sve_svcmplt_wide_s8(__VA_ARGS__) -#define svcmplt_wide_s32(...) __builtin_sve_svcmplt_wide_s32(__VA_ARGS__) -#define svcmplt_wide_s16(...) __builtin_sve_svcmplt_wide_s16(__VA_ARGS__) -#define svcmpne_n_f64(...) __builtin_sve_svcmpne_n_f64(__VA_ARGS__) -#define svcmpne_n_f32(...) __builtin_sve_svcmpne_n_f32(__VA_ARGS__) -#define svcmpne_n_f16(...) __builtin_sve_svcmpne_n_f16(__VA_ARGS__) -#define svcmpne_n_u8(...) __builtin_sve_svcmpne_n_u8(__VA_ARGS__) -#define svcmpne_n_u32(...) __builtin_sve_svcmpne_n_u32(__VA_ARGS__) -#define svcmpne_n_u64(...) __builtin_sve_svcmpne_n_u64(__VA_ARGS__) -#define svcmpne_n_u16(...) __builtin_sve_svcmpne_n_u16(__VA_ARGS__) -#define svcmpne_n_s8(...) __builtin_sve_svcmpne_n_s8(__VA_ARGS__) -#define svcmpne_n_s32(...) __builtin_sve_svcmpne_n_s32(__VA_ARGS__) -#define svcmpne_n_s64(...) __builtin_sve_svcmpne_n_s64(__VA_ARGS__) -#define svcmpne_n_s16(...) __builtin_sve_svcmpne_n_s16(__VA_ARGS__) -#define svcmpne_u8(...) __builtin_sve_svcmpne_u8(__VA_ARGS__) -#define svcmpne_u32(...) __builtin_sve_svcmpne_u32(__VA_ARGS__) -#define svcmpne_u64(...) __builtin_sve_svcmpne_u64(__VA_ARGS__) -#define svcmpne_u16(...) __builtin_sve_svcmpne_u16(__VA_ARGS__) -#define svcmpne_s8(...) __builtin_sve_svcmpne_s8(__VA_ARGS__) -#define svcmpne_s32(...) __builtin_sve_svcmpne_s32(__VA_ARGS__) -#define svcmpne_s64(...) __builtin_sve_svcmpne_s64(__VA_ARGS__) -#define svcmpne_s16(...) __builtin_sve_svcmpne_s16(__VA_ARGS__) -#define svcmpne_f64(...) __builtin_sve_svcmpne_f64(__VA_ARGS__) -#define svcmpne_f32(...) __builtin_sve_svcmpne_f32(__VA_ARGS__) -#define svcmpne_f16(...) __builtin_sve_svcmpne_f16(__VA_ARGS__) -#define svcmpne_wide_n_s8(...) __builtin_sve_svcmpne_wide_n_s8(__VA_ARGS__) -#define svcmpne_wide_n_s32(...) __builtin_sve_svcmpne_wide_n_s32(__VA_ARGS__) -#define svcmpne_wide_n_s16(...) __builtin_sve_svcmpne_wide_n_s16(__VA_ARGS__) -#define svcmpne_wide_s8(...) __builtin_sve_svcmpne_wide_s8(__VA_ARGS__) -#define svcmpne_wide_s32(...) __builtin_sve_svcmpne_wide_s32(__VA_ARGS__) -#define svcmpne_wide_s16(...) __builtin_sve_svcmpne_wide_s16(__VA_ARGS__) -#define svcmpuo_n_f64(...) __builtin_sve_svcmpuo_n_f64(__VA_ARGS__) -#define svcmpuo_n_f32(...) __builtin_sve_svcmpuo_n_f32(__VA_ARGS__) -#define svcmpuo_n_f16(...) __builtin_sve_svcmpuo_n_f16(__VA_ARGS__) -#define svcmpuo_f64(...) __builtin_sve_svcmpuo_f64(__VA_ARGS__) -#define svcmpuo_f32(...) __builtin_sve_svcmpuo_f32(__VA_ARGS__) -#define svcmpuo_f16(...) __builtin_sve_svcmpuo_f16(__VA_ARGS__) -#define svcnot_u8_m(...) __builtin_sve_svcnot_u8_m(__VA_ARGS__) -#define svcnot_u32_m(...) __builtin_sve_svcnot_u32_m(__VA_ARGS__) -#define svcnot_u64_m(...) __builtin_sve_svcnot_u64_m(__VA_ARGS__) -#define svcnot_u16_m(...) __builtin_sve_svcnot_u16_m(__VA_ARGS__) -#define svcnot_s8_m(...) __builtin_sve_svcnot_s8_m(__VA_ARGS__) -#define svcnot_s32_m(...) __builtin_sve_svcnot_s32_m(__VA_ARGS__) -#define svcnot_s64_m(...) __builtin_sve_svcnot_s64_m(__VA_ARGS__) -#define svcnot_s16_m(...) __builtin_sve_svcnot_s16_m(__VA_ARGS__) -#define svcnot_u8_x(...) __builtin_sve_svcnot_u8_x(__VA_ARGS__) -#define svcnot_u32_x(...) __builtin_sve_svcnot_u32_x(__VA_ARGS__) -#define svcnot_u64_x(...) __builtin_sve_svcnot_u64_x(__VA_ARGS__) -#define svcnot_u16_x(...) __builtin_sve_svcnot_u16_x(__VA_ARGS__) -#define svcnot_s8_x(...) __builtin_sve_svcnot_s8_x(__VA_ARGS__) -#define svcnot_s32_x(...) __builtin_sve_svcnot_s32_x(__VA_ARGS__) -#define svcnot_s64_x(...) __builtin_sve_svcnot_s64_x(__VA_ARGS__) -#define svcnot_s16_x(...) __builtin_sve_svcnot_s16_x(__VA_ARGS__) -#define svcnot_u8_z(...) __builtin_sve_svcnot_u8_z(__VA_ARGS__) -#define svcnot_u32_z(...) __builtin_sve_svcnot_u32_z(__VA_ARGS__) -#define svcnot_u64_z(...) __builtin_sve_svcnot_u64_z(__VA_ARGS__) -#define svcnot_u16_z(...) __builtin_sve_svcnot_u16_z(__VA_ARGS__) -#define svcnot_s8_z(...) __builtin_sve_svcnot_s8_z(__VA_ARGS__) -#define svcnot_s32_z(...) __builtin_sve_svcnot_s32_z(__VA_ARGS__) -#define svcnot_s64_z(...) __builtin_sve_svcnot_s64_z(__VA_ARGS__) -#define svcnot_s16_z(...) __builtin_sve_svcnot_s16_z(__VA_ARGS__) -#define svcnt_u8_m(...) __builtin_sve_svcnt_u8_m(__VA_ARGS__) -#define svcnt_u32_m(...) __builtin_sve_svcnt_u32_m(__VA_ARGS__) -#define svcnt_u64_m(...) __builtin_sve_svcnt_u64_m(__VA_ARGS__) -#define svcnt_u16_m(...) __builtin_sve_svcnt_u16_m(__VA_ARGS__) -#define svcnt_s8_m(...) __builtin_sve_svcnt_s8_m(__VA_ARGS__) -#define svcnt_f64_m(...) __builtin_sve_svcnt_f64_m(__VA_ARGS__) -#define svcnt_f32_m(...) __builtin_sve_svcnt_f32_m(__VA_ARGS__) -#define svcnt_f16_m(...) __builtin_sve_svcnt_f16_m(__VA_ARGS__) -#define svcnt_s32_m(...) __builtin_sve_svcnt_s32_m(__VA_ARGS__) -#define svcnt_s64_m(...) __builtin_sve_svcnt_s64_m(__VA_ARGS__) -#define svcnt_s16_m(...) __builtin_sve_svcnt_s16_m(__VA_ARGS__) -#define svcnt_u8_x(...) __builtin_sve_svcnt_u8_x(__VA_ARGS__) -#define svcnt_u32_x(...) __builtin_sve_svcnt_u32_x(__VA_ARGS__) -#define svcnt_u64_x(...) __builtin_sve_svcnt_u64_x(__VA_ARGS__) -#define svcnt_u16_x(...) __builtin_sve_svcnt_u16_x(__VA_ARGS__) -#define svcnt_s8_x(...) __builtin_sve_svcnt_s8_x(__VA_ARGS__) -#define svcnt_f64_x(...) __builtin_sve_svcnt_f64_x(__VA_ARGS__) -#define svcnt_f32_x(...) __builtin_sve_svcnt_f32_x(__VA_ARGS__) -#define svcnt_f16_x(...) __builtin_sve_svcnt_f16_x(__VA_ARGS__) -#define svcnt_s32_x(...) __builtin_sve_svcnt_s32_x(__VA_ARGS__) -#define svcnt_s64_x(...) __builtin_sve_svcnt_s64_x(__VA_ARGS__) -#define svcnt_s16_x(...) __builtin_sve_svcnt_s16_x(__VA_ARGS__) -#define svcnt_u8_z(...) __builtin_sve_svcnt_u8_z(__VA_ARGS__) -#define svcnt_u32_z(...) __builtin_sve_svcnt_u32_z(__VA_ARGS__) -#define svcnt_u64_z(...) __builtin_sve_svcnt_u64_z(__VA_ARGS__) -#define svcnt_u16_z(...) __builtin_sve_svcnt_u16_z(__VA_ARGS__) -#define svcnt_s8_z(...) __builtin_sve_svcnt_s8_z(__VA_ARGS__) -#define svcnt_f64_z(...) __builtin_sve_svcnt_f64_z(__VA_ARGS__) -#define svcnt_f32_z(...) __builtin_sve_svcnt_f32_z(__VA_ARGS__) -#define svcnt_f16_z(...) __builtin_sve_svcnt_f16_z(__VA_ARGS__) -#define svcnt_s32_z(...) __builtin_sve_svcnt_s32_z(__VA_ARGS__) -#define svcnt_s64_z(...) __builtin_sve_svcnt_s64_z(__VA_ARGS__) -#define svcnt_s16_z(...) __builtin_sve_svcnt_s16_z(__VA_ARGS__) -#define svcntb(...) __builtin_sve_svcntb(__VA_ARGS__) -#define svcntb_pat(...) __builtin_sve_svcntb_pat(__VA_ARGS__) -#define svcntd(...) __builtin_sve_svcntd(__VA_ARGS__) -#define svcntd_pat(...) __builtin_sve_svcntd_pat(__VA_ARGS__) -#define svcnth(...) __builtin_sve_svcnth(__VA_ARGS__) -#define svcnth_pat(...) __builtin_sve_svcnth_pat(__VA_ARGS__) -#define svcntp_b8(...) __builtin_sve_svcntp_b8(__VA_ARGS__) -#define svcntp_b32(...) __builtin_sve_svcntp_b32(__VA_ARGS__) -#define svcntp_b64(...) __builtin_sve_svcntp_b64(__VA_ARGS__) -#define svcntp_b16(...) __builtin_sve_svcntp_b16(__VA_ARGS__) -#define svcntw(...) __builtin_sve_svcntw(__VA_ARGS__) -#define svcntw_pat(...) __builtin_sve_svcntw_pat(__VA_ARGS__) -#define svcompact_u32(...) __builtin_sve_svcompact_u32(__VA_ARGS__) -#define svcompact_u64(...) __builtin_sve_svcompact_u64(__VA_ARGS__) -#define svcompact_f64(...) __builtin_sve_svcompact_f64(__VA_ARGS__) -#define svcompact_f32(...) __builtin_sve_svcompact_f32(__VA_ARGS__) -#define svcompact_s32(...) __builtin_sve_svcompact_s32(__VA_ARGS__) -#define svcompact_s64(...) __builtin_sve_svcompact_s64(__VA_ARGS__) -#define svcreate2_u8(...) __builtin_sve_svcreate2_u8(__VA_ARGS__) -#define svcreate2_u32(...) __builtin_sve_svcreate2_u32(__VA_ARGS__) -#define svcreate2_u64(...) __builtin_sve_svcreate2_u64(__VA_ARGS__) -#define svcreate2_u16(...) __builtin_sve_svcreate2_u16(__VA_ARGS__) -#define svcreate2_s8(...) __builtin_sve_svcreate2_s8(__VA_ARGS__) -#define svcreate2_f64(...) __builtin_sve_svcreate2_f64(__VA_ARGS__) -#define svcreate2_f32(...) __builtin_sve_svcreate2_f32(__VA_ARGS__) -#define svcreate2_f16(...) __builtin_sve_svcreate2_f16(__VA_ARGS__) -#define svcreate2_s32(...) __builtin_sve_svcreate2_s32(__VA_ARGS__) -#define svcreate2_s64(...) __builtin_sve_svcreate2_s64(__VA_ARGS__) -#define svcreate2_s16(...) __builtin_sve_svcreate2_s16(__VA_ARGS__) -#define svcreate3_u8(...) __builtin_sve_svcreate3_u8(__VA_ARGS__) -#define svcreate3_u32(...) __builtin_sve_svcreate3_u32(__VA_ARGS__) -#define svcreate3_u64(...) __builtin_sve_svcreate3_u64(__VA_ARGS__) -#define svcreate3_u16(...) __builtin_sve_svcreate3_u16(__VA_ARGS__) -#define svcreate3_s8(...) __builtin_sve_svcreate3_s8(__VA_ARGS__) -#define svcreate3_f64(...) __builtin_sve_svcreate3_f64(__VA_ARGS__) -#define svcreate3_f32(...) __builtin_sve_svcreate3_f32(__VA_ARGS__) -#define svcreate3_f16(...) __builtin_sve_svcreate3_f16(__VA_ARGS__) -#define svcreate3_s32(...) __builtin_sve_svcreate3_s32(__VA_ARGS__) -#define svcreate3_s64(...) __builtin_sve_svcreate3_s64(__VA_ARGS__) -#define svcreate3_s16(...) __builtin_sve_svcreate3_s16(__VA_ARGS__) -#define svcreate4_u8(...) __builtin_sve_svcreate4_u8(__VA_ARGS__) -#define svcreate4_u32(...) __builtin_sve_svcreate4_u32(__VA_ARGS__) -#define svcreate4_u64(...) __builtin_sve_svcreate4_u64(__VA_ARGS__) -#define svcreate4_u16(...) __builtin_sve_svcreate4_u16(__VA_ARGS__) -#define svcreate4_s8(...) __builtin_sve_svcreate4_s8(__VA_ARGS__) -#define svcreate4_f64(...) __builtin_sve_svcreate4_f64(__VA_ARGS__) -#define svcreate4_f32(...) __builtin_sve_svcreate4_f32(__VA_ARGS__) -#define svcreate4_f16(...) __builtin_sve_svcreate4_f16(__VA_ARGS__) -#define svcreate4_s32(...) __builtin_sve_svcreate4_s32(__VA_ARGS__) -#define svcreate4_s64(...) __builtin_sve_svcreate4_s64(__VA_ARGS__) -#define svcreate4_s16(...) __builtin_sve_svcreate4_s16(__VA_ARGS__) -#define svcvt_f16_f32_m(...) __builtin_sve_svcvt_f16_f32_m(__VA_ARGS__) -#define svcvt_f16_f32_x(...) __builtin_sve_svcvt_f16_f32_x(__VA_ARGS__) -#define svcvt_f16_f32_z(...) __builtin_sve_svcvt_f16_f32_z(__VA_ARGS__) -#define svcvt_f16_f64_m(...) __builtin_sve_svcvt_f16_f64_m(__VA_ARGS__) -#define svcvt_f16_f64_x(...) __builtin_sve_svcvt_f16_f64_x(__VA_ARGS__) -#define svcvt_f16_f64_z(...) __builtin_sve_svcvt_f16_f64_z(__VA_ARGS__) -#define svcvt_f16_s16_m(...) __builtin_sve_svcvt_f16_s16_m(__VA_ARGS__) -#define svcvt_f16_s16_x(...) __builtin_sve_svcvt_f16_s16_x(__VA_ARGS__) -#define svcvt_f16_s16_z(...) __builtin_sve_svcvt_f16_s16_z(__VA_ARGS__) -#define svcvt_f16_s32_m(...) __builtin_sve_svcvt_f16_s32_m(__VA_ARGS__) -#define svcvt_f16_s32_x(...) __builtin_sve_svcvt_f16_s32_x(__VA_ARGS__) -#define svcvt_f16_s32_z(...) __builtin_sve_svcvt_f16_s32_z(__VA_ARGS__) -#define svcvt_f16_s64_m(...) __builtin_sve_svcvt_f16_s64_m(__VA_ARGS__) -#define svcvt_f16_s64_x(...) __builtin_sve_svcvt_f16_s64_x(__VA_ARGS__) -#define svcvt_f16_s64_z(...) __builtin_sve_svcvt_f16_s64_z(__VA_ARGS__) -#define svcvt_f16_u16_m(...) __builtin_sve_svcvt_f16_u16_m(__VA_ARGS__) -#define svcvt_f16_u16_x(...) __builtin_sve_svcvt_f16_u16_x(__VA_ARGS__) -#define svcvt_f16_u16_z(...) __builtin_sve_svcvt_f16_u16_z(__VA_ARGS__) -#define svcvt_f16_u32_m(...) __builtin_sve_svcvt_f16_u32_m(__VA_ARGS__) -#define svcvt_f16_u32_x(...) __builtin_sve_svcvt_f16_u32_x(__VA_ARGS__) -#define svcvt_f16_u32_z(...) __builtin_sve_svcvt_f16_u32_z(__VA_ARGS__) -#define svcvt_f16_u64_m(...) __builtin_sve_svcvt_f16_u64_m(__VA_ARGS__) -#define svcvt_f16_u64_x(...) __builtin_sve_svcvt_f16_u64_x(__VA_ARGS__) -#define svcvt_f16_u64_z(...) __builtin_sve_svcvt_f16_u64_z(__VA_ARGS__) -#define svcvt_f32_f16_m(...) __builtin_sve_svcvt_f32_f16_m(__VA_ARGS__) -#define svcvt_f32_f16_x(...) __builtin_sve_svcvt_f32_f16_x(__VA_ARGS__) -#define svcvt_f32_f16_z(...) __builtin_sve_svcvt_f32_f16_z(__VA_ARGS__) -#define svcvt_f32_f64_m(...) __builtin_sve_svcvt_f32_f64_m(__VA_ARGS__) -#define svcvt_f32_f64_x(...) __builtin_sve_svcvt_f32_f64_x(__VA_ARGS__) -#define svcvt_f32_f64_z(...) __builtin_sve_svcvt_f32_f64_z(__VA_ARGS__) -#define svcvt_f32_s32_m(...) __builtin_sve_svcvt_f32_s32_m(__VA_ARGS__) -#define svcvt_f32_s32_x(...) __builtin_sve_svcvt_f32_s32_x(__VA_ARGS__) -#define svcvt_f32_s32_z(...) __builtin_sve_svcvt_f32_s32_z(__VA_ARGS__) -#define svcvt_f32_s64_m(...) __builtin_sve_svcvt_f32_s64_m(__VA_ARGS__) -#define svcvt_f32_s64_x(...) __builtin_sve_svcvt_f32_s64_x(__VA_ARGS__) -#define svcvt_f32_s64_z(...) __builtin_sve_svcvt_f32_s64_z(__VA_ARGS__) -#define svcvt_f32_u32_m(...) __builtin_sve_svcvt_f32_u32_m(__VA_ARGS__) -#define svcvt_f32_u32_x(...) __builtin_sve_svcvt_f32_u32_x(__VA_ARGS__) -#define svcvt_f32_u32_z(...) __builtin_sve_svcvt_f32_u32_z(__VA_ARGS__) -#define svcvt_f32_u64_m(...) __builtin_sve_svcvt_f32_u64_m(__VA_ARGS__) -#define svcvt_f32_u64_x(...) __builtin_sve_svcvt_f32_u64_x(__VA_ARGS__) -#define svcvt_f32_u64_z(...) __builtin_sve_svcvt_f32_u64_z(__VA_ARGS__) -#define svcvt_f64_f16_m(...) __builtin_sve_svcvt_f64_f16_m(__VA_ARGS__) -#define svcvt_f64_f16_x(...) __builtin_sve_svcvt_f64_f16_x(__VA_ARGS__) -#define svcvt_f64_f16_z(...) __builtin_sve_svcvt_f64_f16_z(__VA_ARGS__) -#define svcvt_f64_f32_m(...) __builtin_sve_svcvt_f64_f32_m(__VA_ARGS__) -#define svcvt_f64_f32_x(...) __builtin_sve_svcvt_f64_f32_x(__VA_ARGS__) -#define svcvt_f64_f32_z(...) __builtin_sve_svcvt_f64_f32_z(__VA_ARGS__) -#define svcvt_f64_s32_m(...) __builtin_sve_svcvt_f64_s32_m(__VA_ARGS__) -#define svcvt_f64_s32_x(...) __builtin_sve_svcvt_f64_s32_x(__VA_ARGS__) -#define svcvt_f64_s32_z(...) __builtin_sve_svcvt_f64_s32_z(__VA_ARGS__) -#define svcvt_f64_s64_m(...) __builtin_sve_svcvt_f64_s64_m(__VA_ARGS__) -#define svcvt_f64_s64_x(...) __builtin_sve_svcvt_f64_s64_x(__VA_ARGS__) -#define svcvt_f64_s64_z(...) __builtin_sve_svcvt_f64_s64_z(__VA_ARGS__) -#define svcvt_f64_u32_m(...) __builtin_sve_svcvt_f64_u32_m(__VA_ARGS__) -#define svcvt_f64_u32_x(...) __builtin_sve_svcvt_f64_u32_x(__VA_ARGS__) -#define svcvt_f64_u32_z(...) __builtin_sve_svcvt_f64_u32_z(__VA_ARGS__) -#define svcvt_f64_u64_m(...) __builtin_sve_svcvt_f64_u64_m(__VA_ARGS__) -#define svcvt_f64_u64_x(...) __builtin_sve_svcvt_f64_u64_x(__VA_ARGS__) -#define svcvt_f64_u64_z(...) __builtin_sve_svcvt_f64_u64_z(__VA_ARGS__) -#define svcvt_s16_f16_m(...) __builtin_sve_svcvt_s16_f16_m(__VA_ARGS__) -#define svcvt_s16_f16_x(...) __builtin_sve_svcvt_s16_f16_x(__VA_ARGS__) -#define svcvt_s16_f16_z(...) __builtin_sve_svcvt_s16_f16_z(__VA_ARGS__) -#define svcvt_s32_f16_m(...) __builtin_sve_svcvt_s32_f16_m(__VA_ARGS__) -#define svcvt_s32_f16_x(...) __builtin_sve_svcvt_s32_f16_x(__VA_ARGS__) -#define svcvt_s32_f16_z(...) __builtin_sve_svcvt_s32_f16_z(__VA_ARGS__) -#define svcvt_s32_f32_m(...) __builtin_sve_svcvt_s32_f32_m(__VA_ARGS__) -#define svcvt_s32_f32_x(...) __builtin_sve_svcvt_s32_f32_x(__VA_ARGS__) -#define svcvt_s32_f32_z(...) __builtin_sve_svcvt_s32_f32_z(__VA_ARGS__) -#define svcvt_s32_f64_m(...) __builtin_sve_svcvt_s32_f64_m(__VA_ARGS__) -#define svcvt_s32_f64_x(...) __builtin_sve_svcvt_s32_f64_x(__VA_ARGS__) -#define svcvt_s32_f64_z(...) __builtin_sve_svcvt_s32_f64_z(__VA_ARGS__) -#define svcvt_s64_f16_m(...) __builtin_sve_svcvt_s64_f16_m(__VA_ARGS__) -#define svcvt_s64_f16_x(...) __builtin_sve_svcvt_s64_f16_x(__VA_ARGS__) -#define svcvt_s64_f16_z(...) __builtin_sve_svcvt_s64_f16_z(__VA_ARGS__) -#define svcvt_s64_f32_m(...) __builtin_sve_svcvt_s64_f32_m(__VA_ARGS__) -#define svcvt_s64_f32_x(...) __builtin_sve_svcvt_s64_f32_x(__VA_ARGS__) -#define svcvt_s64_f32_z(...) __builtin_sve_svcvt_s64_f32_z(__VA_ARGS__) -#define svcvt_s64_f64_m(...) __builtin_sve_svcvt_s64_f64_m(__VA_ARGS__) -#define svcvt_s64_f64_x(...) __builtin_sve_svcvt_s64_f64_x(__VA_ARGS__) -#define svcvt_s64_f64_z(...) __builtin_sve_svcvt_s64_f64_z(__VA_ARGS__) -#define svcvt_u16_f16_m(...) __builtin_sve_svcvt_u16_f16_m(__VA_ARGS__) -#define svcvt_u16_f16_x(...) __builtin_sve_svcvt_u16_f16_x(__VA_ARGS__) -#define svcvt_u16_f16_z(...) __builtin_sve_svcvt_u16_f16_z(__VA_ARGS__) -#define svcvt_u32_f16_m(...) __builtin_sve_svcvt_u32_f16_m(__VA_ARGS__) -#define svcvt_u32_f16_x(...) __builtin_sve_svcvt_u32_f16_x(__VA_ARGS__) -#define svcvt_u32_f16_z(...) __builtin_sve_svcvt_u32_f16_z(__VA_ARGS__) -#define svcvt_u32_f32_m(...) __builtin_sve_svcvt_u32_f32_m(__VA_ARGS__) -#define svcvt_u32_f32_x(...) __builtin_sve_svcvt_u32_f32_x(__VA_ARGS__) -#define svcvt_u32_f32_z(...) __builtin_sve_svcvt_u32_f32_z(__VA_ARGS__) -#define svcvt_u32_f64_m(...) __builtin_sve_svcvt_u32_f64_m(__VA_ARGS__) -#define svcvt_u32_f64_x(...) __builtin_sve_svcvt_u32_f64_x(__VA_ARGS__) -#define svcvt_u32_f64_z(...) __builtin_sve_svcvt_u32_f64_z(__VA_ARGS__) -#define svcvt_u64_f16_m(...) __builtin_sve_svcvt_u64_f16_m(__VA_ARGS__) -#define svcvt_u64_f16_x(...) __builtin_sve_svcvt_u64_f16_x(__VA_ARGS__) -#define svcvt_u64_f16_z(...) __builtin_sve_svcvt_u64_f16_z(__VA_ARGS__) -#define svcvt_u64_f32_m(...) __builtin_sve_svcvt_u64_f32_m(__VA_ARGS__) -#define svcvt_u64_f32_x(...) __builtin_sve_svcvt_u64_f32_x(__VA_ARGS__) -#define svcvt_u64_f32_z(...) __builtin_sve_svcvt_u64_f32_z(__VA_ARGS__) -#define svcvt_u64_f64_m(...) __builtin_sve_svcvt_u64_f64_m(__VA_ARGS__) -#define svcvt_u64_f64_x(...) __builtin_sve_svcvt_u64_f64_x(__VA_ARGS__) -#define svcvt_u64_f64_z(...) __builtin_sve_svcvt_u64_f64_z(__VA_ARGS__) -#define svdiv_n_f64_m(...) __builtin_sve_svdiv_n_f64_m(__VA_ARGS__) -#define svdiv_n_f32_m(...) __builtin_sve_svdiv_n_f32_m(__VA_ARGS__) -#define svdiv_n_f16_m(...) __builtin_sve_svdiv_n_f16_m(__VA_ARGS__) -#define svdiv_n_f64_x(...) __builtin_sve_svdiv_n_f64_x(__VA_ARGS__) -#define svdiv_n_f32_x(...) __builtin_sve_svdiv_n_f32_x(__VA_ARGS__) -#define svdiv_n_f16_x(...) __builtin_sve_svdiv_n_f16_x(__VA_ARGS__) -#define svdiv_n_f64_z(...) __builtin_sve_svdiv_n_f64_z(__VA_ARGS__) -#define svdiv_n_f32_z(...) __builtin_sve_svdiv_n_f32_z(__VA_ARGS__) -#define svdiv_n_f16_z(...) __builtin_sve_svdiv_n_f16_z(__VA_ARGS__) -#define svdiv_n_s32_m(...) __builtin_sve_svdiv_n_s32_m(__VA_ARGS__) -#define svdiv_n_s64_m(...) __builtin_sve_svdiv_n_s64_m(__VA_ARGS__) -#define svdiv_n_s32_x(...) __builtin_sve_svdiv_n_s32_x(__VA_ARGS__) -#define svdiv_n_s64_x(...) __builtin_sve_svdiv_n_s64_x(__VA_ARGS__) -#define svdiv_n_s32_z(...) __builtin_sve_svdiv_n_s32_z(__VA_ARGS__) -#define svdiv_n_s64_z(...) __builtin_sve_svdiv_n_s64_z(__VA_ARGS__) -#define svdiv_n_u32_m(...) __builtin_sve_svdiv_n_u32_m(__VA_ARGS__) -#define svdiv_n_u64_m(...) __builtin_sve_svdiv_n_u64_m(__VA_ARGS__) -#define svdiv_n_u32_x(...) __builtin_sve_svdiv_n_u32_x(__VA_ARGS__) -#define svdiv_n_u64_x(...) __builtin_sve_svdiv_n_u64_x(__VA_ARGS__) -#define svdiv_n_u32_z(...) __builtin_sve_svdiv_n_u32_z(__VA_ARGS__) -#define svdiv_n_u64_z(...) __builtin_sve_svdiv_n_u64_z(__VA_ARGS__) -#define svdiv_f64_m(...) __builtin_sve_svdiv_f64_m(__VA_ARGS__) -#define svdiv_f32_m(...) __builtin_sve_svdiv_f32_m(__VA_ARGS__) -#define svdiv_f16_m(...) __builtin_sve_svdiv_f16_m(__VA_ARGS__) -#define svdiv_f64_x(...) __builtin_sve_svdiv_f64_x(__VA_ARGS__) -#define svdiv_f32_x(...) __builtin_sve_svdiv_f32_x(__VA_ARGS__) -#define svdiv_f16_x(...) __builtin_sve_svdiv_f16_x(__VA_ARGS__) -#define svdiv_f64_z(...) __builtin_sve_svdiv_f64_z(__VA_ARGS__) -#define svdiv_f32_z(...) __builtin_sve_svdiv_f32_z(__VA_ARGS__) -#define svdiv_f16_z(...) __builtin_sve_svdiv_f16_z(__VA_ARGS__) -#define svdiv_s32_m(...) __builtin_sve_svdiv_s32_m(__VA_ARGS__) -#define svdiv_s64_m(...) __builtin_sve_svdiv_s64_m(__VA_ARGS__) -#define svdiv_s32_x(...) __builtin_sve_svdiv_s32_x(__VA_ARGS__) -#define svdiv_s64_x(...) __builtin_sve_svdiv_s64_x(__VA_ARGS__) -#define svdiv_s32_z(...) __builtin_sve_svdiv_s32_z(__VA_ARGS__) -#define svdiv_s64_z(...) __builtin_sve_svdiv_s64_z(__VA_ARGS__) -#define svdiv_u32_m(...) __builtin_sve_svdiv_u32_m(__VA_ARGS__) -#define svdiv_u64_m(...) __builtin_sve_svdiv_u64_m(__VA_ARGS__) -#define svdiv_u32_x(...) __builtin_sve_svdiv_u32_x(__VA_ARGS__) -#define svdiv_u64_x(...) __builtin_sve_svdiv_u64_x(__VA_ARGS__) -#define svdiv_u32_z(...) __builtin_sve_svdiv_u32_z(__VA_ARGS__) -#define svdiv_u64_z(...) __builtin_sve_svdiv_u64_z(__VA_ARGS__) -#define svdivr_n_f64_m(...) __builtin_sve_svdivr_n_f64_m(__VA_ARGS__) -#define svdivr_n_f32_m(...) __builtin_sve_svdivr_n_f32_m(__VA_ARGS__) -#define svdivr_n_f16_m(...) __builtin_sve_svdivr_n_f16_m(__VA_ARGS__) -#define svdivr_n_f64_x(...) __builtin_sve_svdivr_n_f64_x(__VA_ARGS__) -#define svdivr_n_f32_x(...) __builtin_sve_svdivr_n_f32_x(__VA_ARGS__) -#define svdivr_n_f16_x(...) __builtin_sve_svdivr_n_f16_x(__VA_ARGS__) -#define svdivr_n_f64_z(...) __builtin_sve_svdivr_n_f64_z(__VA_ARGS__) -#define svdivr_n_f32_z(...) __builtin_sve_svdivr_n_f32_z(__VA_ARGS__) -#define svdivr_n_f16_z(...) __builtin_sve_svdivr_n_f16_z(__VA_ARGS__) -#define svdivr_n_s32_m(...) __builtin_sve_svdivr_n_s32_m(__VA_ARGS__) -#define svdivr_n_s64_m(...) __builtin_sve_svdivr_n_s64_m(__VA_ARGS__) -#define svdivr_n_s32_x(...) __builtin_sve_svdivr_n_s32_x(__VA_ARGS__) -#define svdivr_n_s64_x(...) __builtin_sve_svdivr_n_s64_x(__VA_ARGS__) -#define svdivr_n_s32_z(...) __builtin_sve_svdivr_n_s32_z(__VA_ARGS__) -#define svdivr_n_s64_z(...) __builtin_sve_svdivr_n_s64_z(__VA_ARGS__) -#define svdivr_n_u32_m(...) __builtin_sve_svdivr_n_u32_m(__VA_ARGS__) -#define svdivr_n_u64_m(...) __builtin_sve_svdivr_n_u64_m(__VA_ARGS__) -#define svdivr_n_u32_x(...) __builtin_sve_svdivr_n_u32_x(__VA_ARGS__) -#define svdivr_n_u64_x(...) __builtin_sve_svdivr_n_u64_x(__VA_ARGS__) -#define svdivr_n_u32_z(...) __builtin_sve_svdivr_n_u32_z(__VA_ARGS__) -#define svdivr_n_u64_z(...) __builtin_sve_svdivr_n_u64_z(__VA_ARGS__) -#define svdivr_f64_m(...) __builtin_sve_svdivr_f64_m(__VA_ARGS__) -#define svdivr_f32_m(...) __builtin_sve_svdivr_f32_m(__VA_ARGS__) -#define svdivr_f16_m(...) __builtin_sve_svdivr_f16_m(__VA_ARGS__) -#define svdivr_f64_x(...) __builtin_sve_svdivr_f64_x(__VA_ARGS__) -#define svdivr_f32_x(...) __builtin_sve_svdivr_f32_x(__VA_ARGS__) -#define svdivr_f16_x(...) __builtin_sve_svdivr_f16_x(__VA_ARGS__) -#define svdivr_f64_z(...) __builtin_sve_svdivr_f64_z(__VA_ARGS__) -#define svdivr_f32_z(...) __builtin_sve_svdivr_f32_z(__VA_ARGS__) -#define svdivr_f16_z(...) __builtin_sve_svdivr_f16_z(__VA_ARGS__) -#define svdivr_s32_m(...) __builtin_sve_svdivr_s32_m(__VA_ARGS__) -#define svdivr_s64_m(...) __builtin_sve_svdivr_s64_m(__VA_ARGS__) -#define svdivr_s32_x(...) __builtin_sve_svdivr_s32_x(__VA_ARGS__) -#define svdivr_s64_x(...) __builtin_sve_svdivr_s64_x(__VA_ARGS__) -#define svdivr_s32_z(...) __builtin_sve_svdivr_s32_z(__VA_ARGS__) -#define svdivr_s64_z(...) __builtin_sve_svdivr_s64_z(__VA_ARGS__) -#define svdivr_u32_m(...) __builtin_sve_svdivr_u32_m(__VA_ARGS__) -#define svdivr_u64_m(...) __builtin_sve_svdivr_u64_m(__VA_ARGS__) -#define svdivr_u32_x(...) __builtin_sve_svdivr_u32_x(__VA_ARGS__) -#define svdivr_u64_x(...) __builtin_sve_svdivr_u64_x(__VA_ARGS__) -#define svdivr_u32_z(...) __builtin_sve_svdivr_u32_z(__VA_ARGS__) -#define svdivr_u64_z(...) __builtin_sve_svdivr_u64_z(__VA_ARGS__) -#define svdot_n_s32(...) __builtin_sve_svdot_n_s32(__VA_ARGS__) -#define svdot_n_s64(...) __builtin_sve_svdot_n_s64(__VA_ARGS__) -#define svdot_n_u32(...) __builtin_sve_svdot_n_u32(__VA_ARGS__) -#define svdot_n_u64(...) __builtin_sve_svdot_n_u64(__VA_ARGS__) -#define svdot_s32(...) __builtin_sve_svdot_s32(__VA_ARGS__) -#define svdot_s64(...) __builtin_sve_svdot_s64(__VA_ARGS__) -#define svdot_u32(...) __builtin_sve_svdot_u32(__VA_ARGS__) -#define svdot_u64(...) __builtin_sve_svdot_u64(__VA_ARGS__) -#define svdot_lane_s32(...) __builtin_sve_svdot_lane_s32(__VA_ARGS__) -#define svdot_lane_s64(...) __builtin_sve_svdot_lane_s64(__VA_ARGS__) -#define svdot_lane_u32(...) __builtin_sve_svdot_lane_u32(__VA_ARGS__) -#define svdot_lane_u64(...) __builtin_sve_svdot_lane_u64(__VA_ARGS__) -#define svdup_n_u8(...) __builtin_sve_svdup_n_u8(__VA_ARGS__) -#define svdup_n_u32(...) __builtin_sve_svdup_n_u32(__VA_ARGS__) -#define svdup_n_u64(...) __builtin_sve_svdup_n_u64(__VA_ARGS__) -#define svdup_n_u16(...) __builtin_sve_svdup_n_u16(__VA_ARGS__) -#define svdup_n_s8(...) __builtin_sve_svdup_n_s8(__VA_ARGS__) -#define svdup_n_f64(...) __builtin_sve_svdup_n_f64(__VA_ARGS__) -#define svdup_n_f32(...) __builtin_sve_svdup_n_f32(__VA_ARGS__) -#define svdup_n_f16(...) __builtin_sve_svdup_n_f16(__VA_ARGS__) -#define svdup_n_s32(...) __builtin_sve_svdup_n_s32(__VA_ARGS__) -#define svdup_n_s64(...) __builtin_sve_svdup_n_s64(__VA_ARGS__) -#define svdup_n_s16(...) __builtin_sve_svdup_n_s16(__VA_ARGS__) -#define svdup_n_u8_m(...) __builtin_sve_svdup_n_u8_m(__VA_ARGS__) -#define svdup_n_u32_m(...) __builtin_sve_svdup_n_u32_m(__VA_ARGS__) -#define svdup_n_u64_m(...) __builtin_sve_svdup_n_u64_m(__VA_ARGS__) -#define svdup_n_u16_m(...) __builtin_sve_svdup_n_u16_m(__VA_ARGS__) -#define svdup_n_s8_m(...) __builtin_sve_svdup_n_s8_m(__VA_ARGS__) -#define svdup_n_f64_m(...) __builtin_sve_svdup_n_f64_m(__VA_ARGS__) -#define svdup_n_f32_m(...) __builtin_sve_svdup_n_f32_m(__VA_ARGS__) -#define svdup_n_f16_m(...) __builtin_sve_svdup_n_f16_m(__VA_ARGS__) -#define svdup_n_s32_m(...) __builtin_sve_svdup_n_s32_m(__VA_ARGS__) -#define svdup_n_s64_m(...) __builtin_sve_svdup_n_s64_m(__VA_ARGS__) -#define svdup_n_s16_m(...) __builtin_sve_svdup_n_s16_m(__VA_ARGS__) -#define svdup_n_b8(...) __builtin_sve_svdup_n_b8(__VA_ARGS__) -#define svdup_n_b32(...) __builtin_sve_svdup_n_b32(__VA_ARGS__) -#define svdup_n_b64(...) __builtin_sve_svdup_n_b64(__VA_ARGS__) -#define svdup_n_b16(...) __builtin_sve_svdup_n_b16(__VA_ARGS__) -#define svdup_n_u8_x(...) __builtin_sve_svdup_n_u8_x(__VA_ARGS__) -#define svdup_n_u32_x(...) __builtin_sve_svdup_n_u32_x(__VA_ARGS__) -#define svdup_n_u64_x(...) __builtin_sve_svdup_n_u64_x(__VA_ARGS__) -#define svdup_n_u16_x(...) __builtin_sve_svdup_n_u16_x(__VA_ARGS__) -#define svdup_n_s8_x(...) __builtin_sve_svdup_n_s8_x(__VA_ARGS__) -#define svdup_n_f64_x(...) __builtin_sve_svdup_n_f64_x(__VA_ARGS__) -#define svdup_n_f32_x(...) __builtin_sve_svdup_n_f32_x(__VA_ARGS__) -#define svdup_n_f16_x(...) __builtin_sve_svdup_n_f16_x(__VA_ARGS__) -#define svdup_n_s32_x(...) __builtin_sve_svdup_n_s32_x(__VA_ARGS__) -#define svdup_n_s64_x(...) __builtin_sve_svdup_n_s64_x(__VA_ARGS__) -#define svdup_n_s16_x(...) __builtin_sve_svdup_n_s16_x(__VA_ARGS__) -#define svdup_n_u8_z(...) __builtin_sve_svdup_n_u8_z(__VA_ARGS__) -#define svdup_n_u32_z(...) __builtin_sve_svdup_n_u32_z(__VA_ARGS__) -#define svdup_n_u64_z(...) __builtin_sve_svdup_n_u64_z(__VA_ARGS__) -#define svdup_n_u16_z(...) __builtin_sve_svdup_n_u16_z(__VA_ARGS__) -#define svdup_n_s8_z(...) __builtin_sve_svdup_n_s8_z(__VA_ARGS__) -#define svdup_n_f64_z(...) __builtin_sve_svdup_n_f64_z(__VA_ARGS__) -#define svdup_n_f32_z(...) __builtin_sve_svdup_n_f32_z(__VA_ARGS__) -#define svdup_n_f16_z(...) __builtin_sve_svdup_n_f16_z(__VA_ARGS__) -#define svdup_n_s32_z(...) __builtin_sve_svdup_n_s32_z(__VA_ARGS__) -#define svdup_n_s64_z(...) __builtin_sve_svdup_n_s64_z(__VA_ARGS__) -#define svdup_n_s16_z(...) __builtin_sve_svdup_n_s16_z(__VA_ARGS__) -#define svdup_lane_u8(...) __builtin_sve_svdup_lane_u8(__VA_ARGS__) -#define svdup_lane_u32(...) __builtin_sve_svdup_lane_u32(__VA_ARGS__) -#define svdup_lane_u64(...) __builtin_sve_svdup_lane_u64(__VA_ARGS__) -#define svdup_lane_u16(...) __builtin_sve_svdup_lane_u16(__VA_ARGS__) -#define svdup_lane_s8(...) __builtin_sve_svdup_lane_s8(__VA_ARGS__) -#define svdup_lane_f64(...) __builtin_sve_svdup_lane_f64(__VA_ARGS__) -#define svdup_lane_f32(...) __builtin_sve_svdup_lane_f32(__VA_ARGS__) -#define svdup_lane_f16(...) __builtin_sve_svdup_lane_f16(__VA_ARGS__) -#define svdup_lane_s32(...) __builtin_sve_svdup_lane_s32(__VA_ARGS__) -#define svdup_lane_s64(...) __builtin_sve_svdup_lane_s64(__VA_ARGS__) -#define svdup_lane_s16(...) __builtin_sve_svdup_lane_s16(__VA_ARGS__) -#define svdupq_n_u16(...) __builtin_sve_svdupq_n_u16(__VA_ARGS__) -#define svdupq_n_f16(...) __builtin_sve_svdupq_n_f16(__VA_ARGS__) -#define svdupq_n_s16(...) __builtin_sve_svdupq_n_s16(__VA_ARGS__) -#define svdupq_n_u32(...) __builtin_sve_svdupq_n_u32(__VA_ARGS__) -#define svdupq_n_f32(...) __builtin_sve_svdupq_n_f32(__VA_ARGS__) -#define svdupq_n_s32(...) __builtin_sve_svdupq_n_s32(__VA_ARGS__) -#define svdupq_n_u64(...) __builtin_sve_svdupq_n_u64(__VA_ARGS__) -#define svdupq_n_f64(...) __builtin_sve_svdupq_n_f64(__VA_ARGS__) -#define svdupq_n_s64(...) __builtin_sve_svdupq_n_s64(__VA_ARGS__) -#define svdupq_n_u8(...) __builtin_sve_svdupq_n_u8(__VA_ARGS__) -#define svdupq_n_s8(...) __builtin_sve_svdupq_n_s8(__VA_ARGS__) -#define svdupq_n_b16(...) __builtin_sve_svdupq_n_b16(__VA_ARGS__) -#define svdupq_n_b32(...) __builtin_sve_svdupq_n_b32(__VA_ARGS__) -#define svdupq_n_b64(...) __builtin_sve_svdupq_n_b64(__VA_ARGS__) -#define svdupq_n_b8(...) __builtin_sve_svdupq_n_b8(__VA_ARGS__) -#define svdupq_lane_u8(...) __builtin_sve_svdupq_lane_u8(__VA_ARGS__) -#define svdupq_lane_u32(...) __builtin_sve_svdupq_lane_u32(__VA_ARGS__) -#define svdupq_lane_u64(...) __builtin_sve_svdupq_lane_u64(__VA_ARGS__) -#define svdupq_lane_u16(...) __builtin_sve_svdupq_lane_u16(__VA_ARGS__) -#define svdupq_lane_s8(...) __builtin_sve_svdupq_lane_s8(__VA_ARGS__) -#define svdupq_lane_f64(...) __builtin_sve_svdupq_lane_f64(__VA_ARGS__) -#define svdupq_lane_f32(...) __builtin_sve_svdupq_lane_f32(__VA_ARGS__) -#define svdupq_lane_f16(...) __builtin_sve_svdupq_lane_f16(__VA_ARGS__) -#define svdupq_lane_s32(...) __builtin_sve_svdupq_lane_s32(__VA_ARGS__) -#define svdupq_lane_s64(...) __builtin_sve_svdupq_lane_s64(__VA_ARGS__) -#define svdupq_lane_s16(...) __builtin_sve_svdupq_lane_s16(__VA_ARGS__) -#define sveor_b_z(...) __builtin_sve_sveor_b_z(__VA_ARGS__) -#define sveor_n_u8_m(...) __builtin_sve_sveor_n_u8_m(__VA_ARGS__) -#define sveor_n_u32_m(...) __builtin_sve_sveor_n_u32_m(__VA_ARGS__) -#define sveor_n_u64_m(...) __builtin_sve_sveor_n_u64_m(__VA_ARGS__) -#define sveor_n_u16_m(...) __builtin_sve_sveor_n_u16_m(__VA_ARGS__) -#define sveor_n_s8_m(...) __builtin_sve_sveor_n_s8_m(__VA_ARGS__) -#define sveor_n_s32_m(...) __builtin_sve_sveor_n_s32_m(__VA_ARGS__) -#define sveor_n_s64_m(...) __builtin_sve_sveor_n_s64_m(__VA_ARGS__) -#define sveor_n_s16_m(...) __builtin_sve_sveor_n_s16_m(__VA_ARGS__) -#define sveor_n_u8_x(...) __builtin_sve_sveor_n_u8_x(__VA_ARGS__) -#define sveor_n_u32_x(...) __builtin_sve_sveor_n_u32_x(__VA_ARGS__) -#define sveor_n_u64_x(...) __builtin_sve_sveor_n_u64_x(__VA_ARGS__) -#define sveor_n_u16_x(...) __builtin_sve_sveor_n_u16_x(__VA_ARGS__) -#define sveor_n_s8_x(...) __builtin_sve_sveor_n_s8_x(__VA_ARGS__) -#define sveor_n_s32_x(...) __builtin_sve_sveor_n_s32_x(__VA_ARGS__) -#define sveor_n_s64_x(...) __builtin_sve_sveor_n_s64_x(__VA_ARGS__) -#define sveor_n_s16_x(...) __builtin_sve_sveor_n_s16_x(__VA_ARGS__) -#define sveor_n_u8_z(...) __builtin_sve_sveor_n_u8_z(__VA_ARGS__) -#define sveor_n_u32_z(...) __builtin_sve_sveor_n_u32_z(__VA_ARGS__) -#define sveor_n_u64_z(...) __builtin_sve_sveor_n_u64_z(__VA_ARGS__) -#define sveor_n_u16_z(...) __builtin_sve_sveor_n_u16_z(__VA_ARGS__) -#define sveor_n_s8_z(...) __builtin_sve_sveor_n_s8_z(__VA_ARGS__) -#define sveor_n_s32_z(...) __builtin_sve_sveor_n_s32_z(__VA_ARGS__) -#define sveor_n_s64_z(...) __builtin_sve_sveor_n_s64_z(__VA_ARGS__) -#define sveor_n_s16_z(...) __builtin_sve_sveor_n_s16_z(__VA_ARGS__) -#define sveor_u8_m(...) __builtin_sve_sveor_u8_m(__VA_ARGS__) -#define sveor_u32_m(...) __builtin_sve_sveor_u32_m(__VA_ARGS__) -#define sveor_u64_m(...) __builtin_sve_sveor_u64_m(__VA_ARGS__) -#define sveor_u16_m(...) __builtin_sve_sveor_u16_m(__VA_ARGS__) -#define sveor_s8_m(...) __builtin_sve_sveor_s8_m(__VA_ARGS__) -#define sveor_s32_m(...) __builtin_sve_sveor_s32_m(__VA_ARGS__) -#define sveor_s64_m(...) __builtin_sve_sveor_s64_m(__VA_ARGS__) -#define sveor_s16_m(...) __builtin_sve_sveor_s16_m(__VA_ARGS__) -#define sveor_u8_x(...) __builtin_sve_sveor_u8_x(__VA_ARGS__) -#define sveor_u32_x(...) __builtin_sve_sveor_u32_x(__VA_ARGS__) -#define sveor_u64_x(...) __builtin_sve_sveor_u64_x(__VA_ARGS__) -#define sveor_u16_x(...) __builtin_sve_sveor_u16_x(__VA_ARGS__) -#define sveor_s8_x(...) __builtin_sve_sveor_s8_x(__VA_ARGS__) -#define sveor_s32_x(...) __builtin_sve_sveor_s32_x(__VA_ARGS__) -#define sveor_s64_x(...) __builtin_sve_sveor_s64_x(__VA_ARGS__) -#define sveor_s16_x(...) __builtin_sve_sveor_s16_x(__VA_ARGS__) -#define sveor_u8_z(...) __builtin_sve_sveor_u8_z(__VA_ARGS__) -#define sveor_u32_z(...) __builtin_sve_sveor_u32_z(__VA_ARGS__) -#define sveor_u64_z(...) __builtin_sve_sveor_u64_z(__VA_ARGS__) -#define sveor_u16_z(...) __builtin_sve_sveor_u16_z(__VA_ARGS__) -#define sveor_s8_z(...) __builtin_sve_sveor_s8_z(__VA_ARGS__) -#define sveor_s32_z(...) __builtin_sve_sveor_s32_z(__VA_ARGS__) -#define sveor_s64_z(...) __builtin_sve_sveor_s64_z(__VA_ARGS__) -#define sveor_s16_z(...) __builtin_sve_sveor_s16_z(__VA_ARGS__) -#define sveorv_u8(...) __builtin_sve_sveorv_u8(__VA_ARGS__) -#define sveorv_u32(...) __builtin_sve_sveorv_u32(__VA_ARGS__) -#define sveorv_u64(...) __builtin_sve_sveorv_u64(__VA_ARGS__) -#define sveorv_u16(...) __builtin_sve_sveorv_u16(__VA_ARGS__) -#define sveorv_s8(...) __builtin_sve_sveorv_s8(__VA_ARGS__) -#define sveorv_s32(...) __builtin_sve_sveorv_s32(__VA_ARGS__) -#define sveorv_s64(...) __builtin_sve_sveorv_s64(__VA_ARGS__) -#define sveorv_s16(...) __builtin_sve_sveorv_s16(__VA_ARGS__) -#define svexpa_f64(...) __builtin_sve_svexpa_f64(__VA_ARGS__) -#define svexpa_f32(...) __builtin_sve_svexpa_f32(__VA_ARGS__) -#define svexpa_f16(...) __builtin_sve_svexpa_f16(__VA_ARGS__) -#define svext_u8(...) __builtin_sve_svext_u8(__VA_ARGS__) -#define svext_u32(...) __builtin_sve_svext_u32(__VA_ARGS__) -#define svext_u64(...) __builtin_sve_svext_u64(__VA_ARGS__) -#define svext_u16(...) __builtin_sve_svext_u16(__VA_ARGS__) -#define svext_s8(...) __builtin_sve_svext_s8(__VA_ARGS__) -#define svext_f64(...) __builtin_sve_svext_f64(__VA_ARGS__) -#define svext_f32(...) __builtin_sve_svext_f32(__VA_ARGS__) -#define svext_f16(...) __builtin_sve_svext_f16(__VA_ARGS__) -#define svext_s32(...) __builtin_sve_svext_s32(__VA_ARGS__) -#define svext_s64(...) __builtin_sve_svext_s64(__VA_ARGS__) -#define svext_s16(...) __builtin_sve_svext_s16(__VA_ARGS__) -#define svextb_s32_m(...) __builtin_sve_svextb_s32_m(__VA_ARGS__) -#define svextb_s64_m(...) __builtin_sve_svextb_s64_m(__VA_ARGS__) -#define svextb_s16_m(...) __builtin_sve_svextb_s16_m(__VA_ARGS__) -#define svextb_s32_x(...) __builtin_sve_svextb_s32_x(__VA_ARGS__) -#define svextb_s64_x(...) __builtin_sve_svextb_s64_x(__VA_ARGS__) -#define svextb_s16_x(...) __builtin_sve_svextb_s16_x(__VA_ARGS__) -#define svextb_s32_z(...) __builtin_sve_svextb_s32_z(__VA_ARGS__) -#define svextb_s64_z(...) __builtin_sve_svextb_s64_z(__VA_ARGS__) -#define svextb_s16_z(...) __builtin_sve_svextb_s16_z(__VA_ARGS__) -#define svextb_u32_m(...) __builtin_sve_svextb_u32_m(__VA_ARGS__) -#define svextb_u64_m(...) __builtin_sve_svextb_u64_m(__VA_ARGS__) -#define svextb_u16_m(...) __builtin_sve_svextb_u16_m(__VA_ARGS__) -#define svextb_u32_x(...) __builtin_sve_svextb_u32_x(__VA_ARGS__) -#define svextb_u64_x(...) __builtin_sve_svextb_u64_x(__VA_ARGS__) -#define svextb_u16_x(...) __builtin_sve_svextb_u16_x(__VA_ARGS__) -#define svextb_u32_z(...) __builtin_sve_svextb_u32_z(__VA_ARGS__) -#define svextb_u64_z(...) __builtin_sve_svextb_u64_z(__VA_ARGS__) -#define svextb_u16_z(...) __builtin_sve_svextb_u16_z(__VA_ARGS__) -#define svexth_s32_m(...) __builtin_sve_svexth_s32_m(__VA_ARGS__) -#define svexth_s64_m(...) __builtin_sve_svexth_s64_m(__VA_ARGS__) -#define svexth_s32_x(...) __builtin_sve_svexth_s32_x(__VA_ARGS__) -#define svexth_s64_x(...) __builtin_sve_svexth_s64_x(__VA_ARGS__) -#define svexth_s32_z(...) __builtin_sve_svexth_s32_z(__VA_ARGS__) -#define svexth_s64_z(...) __builtin_sve_svexth_s64_z(__VA_ARGS__) -#define svexth_u32_m(...) __builtin_sve_svexth_u32_m(__VA_ARGS__) -#define svexth_u64_m(...) __builtin_sve_svexth_u64_m(__VA_ARGS__) -#define svexth_u32_x(...) __builtin_sve_svexth_u32_x(__VA_ARGS__) -#define svexth_u64_x(...) __builtin_sve_svexth_u64_x(__VA_ARGS__) -#define svexth_u32_z(...) __builtin_sve_svexth_u32_z(__VA_ARGS__) -#define svexth_u64_z(...) __builtin_sve_svexth_u64_z(__VA_ARGS__) -#define svextw_s64_m(...) __builtin_sve_svextw_s64_m(__VA_ARGS__) -#define svextw_s64_x(...) __builtin_sve_svextw_s64_x(__VA_ARGS__) -#define svextw_s64_z(...) __builtin_sve_svextw_s64_z(__VA_ARGS__) -#define svextw_u64_m(...) __builtin_sve_svextw_u64_m(__VA_ARGS__) -#define svextw_u64_x(...) __builtin_sve_svextw_u64_x(__VA_ARGS__) -#define svextw_u64_z(...) __builtin_sve_svextw_u64_z(__VA_ARGS__) -#define svget2_u8(...) __builtin_sve_svget2_u8(__VA_ARGS__) -#define svget2_u32(...) __builtin_sve_svget2_u32(__VA_ARGS__) -#define svget2_u64(...) __builtin_sve_svget2_u64(__VA_ARGS__) -#define svget2_u16(...) __builtin_sve_svget2_u16(__VA_ARGS__) -#define svget2_s8(...) __builtin_sve_svget2_s8(__VA_ARGS__) -#define svget2_f64(...) __builtin_sve_svget2_f64(__VA_ARGS__) -#define svget2_f32(...) __builtin_sve_svget2_f32(__VA_ARGS__) -#define svget2_f16(...) __builtin_sve_svget2_f16(__VA_ARGS__) -#define svget2_s32(...) __builtin_sve_svget2_s32(__VA_ARGS__) -#define svget2_s64(...) __builtin_sve_svget2_s64(__VA_ARGS__) -#define svget2_s16(...) __builtin_sve_svget2_s16(__VA_ARGS__) -#define svget3_u8(...) __builtin_sve_svget3_u8(__VA_ARGS__) -#define svget3_u32(...) __builtin_sve_svget3_u32(__VA_ARGS__) -#define svget3_u64(...) __builtin_sve_svget3_u64(__VA_ARGS__) -#define svget3_u16(...) __builtin_sve_svget3_u16(__VA_ARGS__) -#define svget3_s8(...) __builtin_sve_svget3_s8(__VA_ARGS__) -#define svget3_f64(...) __builtin_sve_svget3_f64(__VA_ARGS__) -#define svget3_f32(...) __builtin_sve_svget3_f32(__VA_ARGS__) -#define svget3_f16(...) __builtin_sve_svget3_f16(__VA_ARGS__) -#define svget3_s32(...) __builtin_sve_svget3_s32(__VA_ARGS__) -#define svget3_s64(...) __builtin_sve_svget3_s64(__VA_ARGS__) -#define svget3_s16(...) __builtin_sve_svget3_s16(__VA_ARGS__) -#define svget4_u8(...) __builtin_sve_svget4_u8(__VA_ARGS__) -#define svget4_u32(...) __builtin_sve_svget4_u32(__VA_ARGS__) -#define svget4_u64(...) __builtin_sve_svget4_u64(__VA_ARGS__) -#define svget4_u16(...) __builtin_sve_svget4_u16(__VA_ARGS__) -#define svget4_s8(...) __builtin_sve_svget4_s8(__VA_ARGS__) -#define svget4_f64(...) __builtin_sve_svget4_f64(__VA_ARGS__) -#define svget4_f32(...) __builtin_sve_svget4_f32(__VA_ARGS__) -#define svget4_f16(...) __builtin_sve_svget4_f16(__VA_ARGS__) -#define svget4_s32(...) __builtin_sve_svget4_s32(__VA_ARGS__) -#define svget4_s64(...) __builtin_sve_svget4_s64(__VA_ARGS__) -#define svget4_s16(...) __builtin_sve_svget4_s16(__VA_ARGS__) -#define svindex_u8(...) __builtin_sve_svindex_u8(__VA_ARGS__) -#define svindex_u32(...) __builtin_sve_svindex_u32(__VA_ARGS__) -#define svindex_u64(...) __builtin_sve_svindex_u64(__VA_ARGS__) -#define svindex_u16(...) __builtin_sve_svindex_u16(__VA_ARGS__) -#define svindex_s8(...) __builtin_sve_svindex_s8(__VA_ARGS__) -#define svindex_s32(...) __builtin_sve_svindex_s32(__VA_ARGS__) -#define svindex_s64(...) __builtin_sve_svindex_s64(__VA_ARGS__) -#define svindex_s16(...) __builtin_sve_svindex_s16(__VA_ARGS__) -#define svinsr_n_u8(...) __builtin_sve_svinsr_n_u8(__VA_ARGS__) -#define svinsr_n_u32(...) __builtin_sve_svinsr_n_u32(__VA_ARGS__) -#define svinsr_n_u64(...) __builtin_sve_svinsr_n_u64(__VA_ARGS__) -#define svinsr_n_u16(...) __builtin_sve_svinsr_n_u16(__VA_ARGS__) -#define svinsr_n_s8(...) __builtin_sve_svinsr_n_s8(__VA_ARGS__) -#define svinsr_n_f64(...) __builtin_sve_svinsr_n_f64(__VA_ARGS__) -#define svinsr_n_f32(...) __builtin_sve_svinsr_n_f32(__VA_ARGS__) -#define svinsr_n_f16(...) __builtin_sve_svinsr_n_f16(__VA_ARGS__) -#define svinsr_n_s32(...) __builtin_sve_svinsr_n_s32(__VA_ARGS__) -#define svinsr_n_s64(...) __builtin_sve_svinsr_n_s64(__VA_ARGS__) -#define svinsr_n_s16(...) __builtin_sve_svinsr_n_s16(__VA_ARGS__) -#define svlasta_u8(...) __builtin_sve_svlasta_u8(__VA_ARGS__) -#define svlasta_u32(...) __builtin_sve_svlasta_u32(__VA_ARGS__) -#define svlasta_u64(...) __builtin_sve_svlasta_u64(__VA_ARGS__) -#define svlasta_u16(...) __builtin_sve_svlasta_u16(__VA_ARGS__) -#define svlasta_s8(...) __builtin_sve_svlasta_s8(__VA_ARGS__) -#define svlasta_f64(...) __builtin_sve_svlasta_f64(__VA_ARGS__) -#define svlasta_f32(...) __builtin_sve_svlasta_f32(__VA_ARGS__) -#define svlasta_f16(...) __builtin_sve_svlasta_f16(__VA_ARGS__) -#define svlasta_s32(...) __builtin_sve_svlasta_s32(__VA_ARGS__) -#define svlasta_s64(...) __builtin_sve_svlasta_s64(__VA_ARGS__) -#define svlasta_s16(...) __builtin_sve_svlasta_s16(__VA_ARGS__) -#define svlastb_u8(...) __builtin_sve_svlastb_u8(__VA_ARGS__) -#define svlastb_u32(...) __builtin_sve_svlastb_u32(__VA_ARGS__) -#define svlastb_u64(...) __builtin_sve_svlastb_u64(__VA_ARGS__) -#define svlastb_u16(...) __builtin_sve_svlastb_u16(__VA_ARGS__) -#define svlastb_s8(...) __builtin_sve_svlastb_s8(__VA_ARGS__) -#define svlastb_f64(...) __builtin_sve_svlastb_f64(__VA_ARGS__) -#define svlastb_f32(...) __builtin_sve_svlastb_f32(__VA_ARGS__) -#define svlastb_f16(...) __builtin_sve_svlastb_f16(__VA_ARGS__) -#define svlastb_s32(...) __builtin_sve_svlastb_s32(__VA_ARGS__) -#define svlastb_s64(...) __builtin_sve_svlastb_s64(__VA_ARGS__) -#define svlastb_s16(...) __builtin_sve_svlastb_s16(__VA_ARGS__) -#define svld1_u8(...) __builtin_sve_svld1_u8(__VA_ARGS__) -#define svld1_u32(...) __builtin_sve_svld1_u32(__VA_ARGS__) -#define svld1_u64(...) __builtin_sve_svld1_u64(__VA_ARGS__) -#define svld1_u16(...) __builtin_sve_svld1_u16(__VA_ARGS__) -#define svld1_s8(...) __builtin_sve_svld1_s8(__VA_ARGS__) -#define svld1_f64(...) __builtin_sve_svld1_f64(__VA_ARGS__) -#define svld1_f32(...) __builtin_sve_svld1_f32(__VA_ARGS__) -#define svld1_f16(...) __builtin_sve_svld1_f16(__VA_ARGS__) -#define svld1_s32(...) __builtin_sve_svld1_s32(__VA_ARGS__) -#define svld1_s64(...) __builtin_sve_svld1_s64(__VA_ARGS__) -#define svld1_s16(...) __builtin_sve_svld1_s16(__VA_ARGS__) -#define svld1_gather_u32base_index_u32(...) __builtin_sve_svld1_gather_u32base_index_u32(__VA_ARGS__) -#define svld1_gather_u64base_index_u64(...) __builtin_sve_svld1_gather_u64base_index_u64(__VA_ARGS__) -#define svld1_gather_u64base_index_f64(...) __builtin_sve_svld1_gather_u64base_index_f64(__VA_ARGS__) -#define svld1_gather_u32base_index_f32(...) __builtin_sve_svld1_gather_u32base_index_f32(__VA_ARGS__) -#define svld1_gather_u32base_index_s32(...) __builtin_sve_svld1_gather_u32base_index_s32(__VA_ARGS__) -#define svld1_gather_u64base_index_s64(...) __builtin_sve_svld1_gather_u64base_index_s64(__VA_ARGS__) -#define svld1_gather_u32base_offset_u32(...) __builtin_sve_svld1_gather_u32base_offset_u32(__VA_ARGS__) -#define svld1_gather_u64base_offset_u64(...) __builtin_sve_svld1_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1_gather_u64base_offset_f64(...) __builtin_sve_svld1_gather_u64base_offset_f64(__VA_ARGS__) -#define svld1_gather_u32base_offset_f32(...) __builtin_sve_svld1_gather_u32base_offset_f32(__VA_ARGS__) -#define svld1_gather_u32base_offset_s32(...) __builtin_sve_svld1_gather_u32base_offset_s32(__VA_ARGS__) -#define svld1_gather_u64base_offset_s64(...) __builtin_sve_svld1_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1_gather_u32base_u32(...) __builtin_sve_svld1_gather_u32base_u32(__VA_ARGS__) -#define svld1_gather_u64base_u64(...) __builtin_sve_svld1_gather_u64base_u64(__VA_ARGS__) -#define svld1_gather_u64base_f64(...) __builtin_sve_svld1_gather_u64base_f64(__VA_ARGS__) -#define svld1_gather_u32base_f32(...) __builtin_sve_svld1_gather_u32base_f32(__VA_ARGS__) -#define svld1_gather_u32base_s32(...) __builtin_sve_svld1_gather_u32base_s32(__VA_ARGS__) -#define svld1_gather_u64base_s64(...) __builtin_sve_svld1_gather_u64base_s64(__VA_ARGS__) -#define svld1_gather_s32index_u32(...) __builtin_sve_svld1_gather_s32index_u32(__VA_ARGS__) -#define svld1_gather_s32index_f32(...) __builtin_sve_svld1_gather_s32index_f32(__VA_ARGS__) -#define svld1_gather_s32index_s32(...) __builtin_sve_svld1_gather_s32index_s32(__VA_ARGS__) -#define svld1_gather_u32index_u32(...) __builtin_sve_svld1_gather_u32index_u32(__VA_ARGS__) -#define svld1_gather_u32index_f32(...) __builtin_sve_svld1_gather_u32index_f32(__VA_ARGS__) -#define svld1_gather_u32index_s32(...) __builtin_sve_svld1_gather_u32index_s32(__VA_ARGS__) -#define svld1_gather_s64index_u64(...) __builtin_sve_svld1_gather_s64index_u64(__VA_ARGS__) -#define svld1_gather_s64index_f64(...) __builtin_sve_svld1_gather_s64index_f64(__VA_ARGS__) -#define svld1_gather_s64index_s64(...) __builtin_sve_svld1_gather_s64index_s64(__VA_ARGS__) -#define svld1_gather_u64index_u64(...) __builtin_sve_svld1_gather_u64index_u64(__VA_ARGS__) -#define svld1_gather_u64index_f64(...) __builtin_sve_svld1_gather_u64index_f64(__VA_ARGS__) -#define svld1_gather_u64index_s64(...) __builtin_sve_svld1_gather_u64index_s64(__VA_ARGS__) -#define svld1_gather_s32offset_u32(...) __builtin_sve_svld1_gather_s32offset_u32(__VA_ARGS__) -#define svld1_gather_s32offset_f32(...) __builtin_sve_svld1_gather_s32offset_f32(__VA_ARGS__) -#define svld1_gather_s32offset_s32(...) __builtin_sve_svld1_gather_s32offset_s32(__VA_ARGS__) -#define svld1_gather_u32offset_u32(...) __builtin_sve_svld1_gather_u32offset_u32(__VA_ARGS__) -#define svld1_gather_u32offset_f32(...) __builtin_sve_svld1_gather_u32offset_f32(__VA_ARGS__) -#define svld1_gather_u32offset_s32(...) __builtin_sve_svld1_gather_u32offset_s32(__VA_ARGS__) -#define svld1_gather_s64offset_u64(...) __builtin_sve_svld1_gather_s64offset_u64(__VA_ARGS__) -#define svld1_gather_s64offset_f64(...) __builtin_sve_svld1_gather_s64offset_f64(__VA_ARGS__) -#define svld1_gather_s64offset_s64(...) __builtin_sve_svld1_gather_s64offset_s64(__VA_ARGS__) -#define svld1_gather_u64offset_u64(...) __builtin_sve_svld1_gather_u64offset_u64(__VA_ARGS__) -#define svld1_gather_u64offset_f64(...) __builtin_sve_svld1_gather_u64offset_f64(__VA_ARGS__) -#define svld1_gather_u64offset_s64(...) __builtin_sve_svld1_gather_u64offset_s64(__VA_ARGS__) -#define svld1_vnum_u8(...) __builtin_sve_svld1_vnum_u8(__VA_ARGS__) -#define svld1_vnum_u32(...) __builtin_sve_svld1_vnum_u32(__VA_ARGS__) -#define svld1_vnum_u64(...) __builtin_sve_svld1_vnum_u64(__VA_ARGS__) -#define svld1_vnum_u16(...) __builtin_sve_svld1_vnum_u16(__VA_ARGS__) -#define svld1_vnum_s8(...) __builtin_sve_svld1_vnum_s8(__VA_ARGS__) -#define svld1_vnum_f64(...) __builtin_sve_svld1_vnum_f64(__VA_ARGS__) -#define svld1_vnum_f32(...) __builtin_sve_svld1_vnum_f32(__VA_ARGS__) -#define svld1_vnum_f16(...) __builtin_sve_svld1_vnum_f16(__VA_ARGS__) -#define svld1_vnum_s32(...) __builtin_sve_svld1_vnum_s32(__VA_ARGS__) -#define svld1_vnum_s64(...) __builtin_sve_svld1_vnum_s64(__VA_ARGS__) -#define svld1_vnum_s16(...) __builtin_sve_svld1_vnum_s16(__VA_ARGS__) -#define svld1rq_u8(...) __builtin_sve_svld1rq_u8(__VA_ARGS__) -#define svld1rq_u32(...) __builtin_sve_svld1rq_u32(__VA_ARGS__) -#define svld1rq_u64(...) __builtin_sve_svld1rq_u64(__VA_ARGS__) -#define svld1rq_u16(...) __builtin_sve_svld1rq_u16(__VA_ARGS__) -#define svld1rq_s8(...) __builtin_sve_svld1rq_s8(__VA_ARGS__) -#define svld1rq_f64(...) __builtin_sve_svld1rq_f64(__VA_ARGS__) -#define svld1rq_f32(...) __builtin_sve_svld1rq_f32(__VA_ARGS__) -#define svld1rq_f16(...) __builtin_sve_svld1rq_f16(__VA_ARGS__) -#define svld1rq_s32(...) __builtin_sve_svld1rq_s32(__VA_ARGS__) -#define svld1rq_s64(...) __builtin_sve_svld1rq_s64(__VA_ARGS__) -#define svld1rq_s16(...) __builtin_sve_svld1rq_s16(__VA_ARGS__) -#define svld1sb_gather_u32base_offset_u32(...) __builtin_sve_svld1sb_gather_u32base_offset_u32(__VA_ARGS__) -#define svld1sb_gather_u64base_offset_u64(...) __builtin_sve_svld1sb_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1sb_gather_u32base_offset_s32(...) __builtin_sve_svld1sb_gather_u32base_offset_s32(__VA_ARGS__) -#define svld1sb_gather_u64base_offset_s64(...) __builtin_sve_svld1sb_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1sb_gather_u32base_u32(...) __builtin_sve_svld1sb_gather_u32base_u32(__VA_ARGS__) -#define svld1sb_gather_u64base_u64(...) __builtin_sve_svld1sb_gather_u64base_u64(__VA_ARGS__) -#define svld1sb_gather_u32base_s32(...) __builtin_sve_svld1sb_gather_u32base_s32(__VA_ARGS__) -#define svld1sb_gather_u64base_s64(...) __builtin_sve_svld1sb_gather_u64base_s64(__VA_ARGS__) -#define svld1sb_gather_s32offset_u32(...) __builtin_sve_svld1sb_gather_s32offset_u32(__VA_ARGS__) -#define svld1sb_gather_s32offset_s32(...) __builtin_sve_svld1sb_gather_s32offset_s32(__VA_ARGS__) -#define svld1sb_gather_u32offset_u32(...) __builtin_sve_svld1sb_gather_u32offset_u32(__VA_ARGS__) -#define svld1sb_gather_u32offset_s32(...) __builtin_sve_svld1sb_gather_u32offset_s32(__VA_ARGS__) -#define svld1sb_gather_s64offset_u64(...) __builtin_sve_svld1sb_gather_s64offset_u64(__VA_ARGS__) -#define svld1sb_gather_s64offset_s64(...) __builtin_sve_svld1sb_gather_s64offset_s64(__VA_ARGS__) -#define svld1sb_gather_u64offset_u64(...) __builtin_sve_svld1sb_gather_u64offset_u64(__VA_ARGS__) -#define svld1sb_gather_u64offset_s64(...) __builtin_sve_svld1sb_gather_u64offset_s64(__VA_ARGS__) -#define svld1sb_vnum_u32(...) __builtin_sve_svld1sb_vnum_u32(__VA_ARGS__) -#define svld1sb_vnum_u64(...) __builtin_sve_svld1sb_vnum_u64(__VA_ARGS__) -#define svld1sb_vnum_u16(...) __builtin_sve_svld1sb_vnum_u16(__VA_ARGS__) -#define svld1sb_vnum_s32(...) __builtin_sve_svld1sb_vnum_s32(__VA_ARGS__) -#define svld1sb_vnum_s64(...) __builtin_sve_svld1sb_vnum_s64(__VA_ARGS__) -#define svld1sb_vnum_s16(...) __builtin_sve_svld1sb_vnum_s16(__VA_ARGS__) -#define svld1sb_u32(...) __builtin_sve_svld1sb_u32(__VA_ARGS__) -#define svld1sb_u64(...) __builtin_sve_svld1sb_u64(__VA_ARGS__) -#define svld1sb_u16(...) __builtin_sve_svld1sb_u16(__VA_ARGS__) -#define svld1sb_s32(...) __builtin_sve_svld1sb_s32(__VA_ARGS__) -#define svld1sb_s64(...) __builtin_sve_svld1sb_s64(__VA_ARGS__) -#define svld1sb_s16(...) __builtin_sve_svld1sb_s16(__VA_ARGS__) -#define svld1sh_gather_u32base_index_u32(...) __builtin_sve_svld1sh_gather_u32base_index_u32(__VA_ARGS__) -#define svld1sh_gather_u64base_index_u64(...) __builtin_sve_svld1sh_gather_u64base_index_u64(__VA_ARGS__) -#define svld1sh_gather_u32base_index_s32(...) __builtin_sve_svld1sh_gather_u32base_index_s32(__VA_ARGS__) -#define svld1sh_gather_u64base_index_s64(...) __builtin_sve_svld1sh_gather_u64base_index_s64(__VA_ARGS__) -#define svld1sh_gather_u32base_offset_u32(...) __builtin_sve_svld1sh_gather_u32base_offset_u32(__VA_ARGS__) -#define svld1sh_gather_u64base_offset_u64(...) __builtin_sve_svld1sh_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1sh_gather_u32base_offset_s32(...) __builtin_sve_svld1sh_gather_u32base_offset_s32(__VA_ARGS__) -#define svld1sh_gather_u64base_offset_s64(...) __builtin_sve_svld1sh_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1sh_gather_u32base_u32(...) __builtin_sve_svld1sh_gather_u32base_u32(__VA_ARGS__) -#define svld1sh_gather_u64base_u64(...) __builtin_sve_svld1sh_gather_u64base_u64(__VA_ARGS__) -#define svld1sh_gather_u32base_s32(...) __builtin_sve_svld1sh_gather_u32base_s32(__VA_ARGS__) -#define svld1sh_gather_u64base_s64(...) __builtin_sve_svld1sh_gather_u64base_s64(__VA_ARGS__) -#define svld1sh_gather_s32index_u32(...) __builtin_sve_svld1sh_gather_s32index_u32(__VA_ARGS__) -#define svld1sh_gather_s32index_s32(...) __builtin_sve_svld1sh_gather_s32index_s32(__VA_ARGS__) -#define svld1sh_gather_u32index_u32(...) __builtin_sve_svld1sh_gather_u32index_u32(__VA_ARGS__) -#define svld1sh_gather_u32index_s32(...) __builtin_sve_svld1sh_gather_u32index_s32(__VA_ARGS__) -#define svld1sh_gather_s64index_u64(...) __builtin_sve_svld1sh_gather_s64index_u64(__VA_ARGS__) -#define svld1sh_gather_s64index_s64(...) __builtin_sve_svld1sh_gather_s64index_s64(__VA_ARGS__) -#define svld1sh_gather_u64index_u64(...) __builtin_sve_svld1sh_gather_u64index_u64(__VA_ARGS__) -#define svld1sh_gather_u64index_s64(...) __builtin_sve_svld1sh_gather_u64index_s64(__VA_ARGS__) -#define svld1sh_gather_s32offset_u32(...) __builtin_sve_svld1sh_gather_s32offset_u32(__VA_ARGS__) -#define svld1sh_gather_s32offset_s32(...) __builtin_sve_svld1sh_gather_s32offset_s32(__VA_ARGS__) -#define svld1sh_gather_u32offset_u32(...) __builtin_sve_svld1sh_gather_u32offset_u32(__VA_ARGS__) -#define svld1sh_gather_u32offset_s32(...) __builtin_sve_svld1sh_gather_u32offset_s32(__VA_ARGS__) -#define svld1sh_gather_s64offset_u64(...) __builtin_sve_svld1sh_gather_s64offset_u64(__VA_ARGS__) -#define svld1sh_gather_s64offset_s64(...) __builtin_sve_svld1sh_gather_s64offset_s64(__VA_ARGS__) -#define svld1sh_gather_u64offset_u64(...) __builtin_sve_svld1sh_gather_u64offset_u64(__VA_ARGS__) -#define svld1sh_gather_u64offset_s64(...) __builtin_sve_svld1sh_gather_u64offset_s64(__VA_ARGS__) -#define svld1sh_vnum_u32(...) __builtin_sve_svld1sh_vnum_u32(__VA_ARGS__) -#define svld1sh_vnum_u64(...) __builtin_sve_svld1sh_vnum_u64(__VA_ARGS__) -#define svld1sh_vnum_s32(...) __builtin_sve_svld1sh_vnum_s32(__VA_ARGS__) -#define svld1sh_vnum_s64(...) __builtin_sve_svld1sh_vnum_s64(__VA_ARGS__) -#define svld1sh_u32(...) __builtin_sve_svld1sh_u32(__VA_ARGS__) -#define svld1sh_u64(...) __builtin_sve_svld1sh_u64(__VA_ARGS__) -#define svld1sh_s32(...) __builtin_sve_svld1sh_s32(__VA_ARGS__) -#define svld1sh_s64(...) __builtin_sve_svld1sh_s64(__VA_ARGS__) -#define svld1sw_gather_u64base_index_u64(...) __builtin_sve_svld1sw_gather_u64base_index_u64(__VA_ARGS__) -#define svld1sw_gather_u64base_index_s64(...) __builtin_sve_svld1sw_gather_u64base_index_s64(__VA_ARGS__) -#define svld1sw_gather_u64base_offset_u64(...) __builtin_sve_svld1sw_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1sw_gather_u64base_offset_s64(...) __builtin_sve_svld1sw_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1sw_gather_u64base_u64(...) __builtin_sve_svld1sw_gather_u64base_u64(__VA_ARGS__) -#define svld1sw_gather_u64base_s64(...) __builtin_sve_svld1sw_gather_u64base_s64(__VA_ARGS__) -#define svld1sw_gather_s64index_u64(...) __builtin_sve_svld1sw_gather_s64index_u64(__VA_ARGS__) -#define svld1sw_gather_s64index_s64(...) __builtin_sve_svld1sw_gather_s64index_s64(__VA_ARGS__) -#define svld1sw_gather_u64index_u64(...) __builtin_sve_svld1sw_gather_u64index_u64(__VA_ARGS__) -#define svld1sw_gather_u64index_s64(...) __builtin_sve_svld1sw_gather_u64index_s64(__VA_ARGS__) -#define svld1sw_gather_s64offset_u64(...) __builtin_sve_svld1sw_gather_s64offset_u64(__VA_ARGS__) -#define svld1sw_gather_s64offset_s64(...) __builtin_sve_svld1sw_gather_s64offset_s64(__VA_ARGS__) -#define svld1sw_gather_u64offset_u64(...) __builtin_sve_svld1sw_gather_u64offset_u64(__VA_ARGS__) -#define svld1sw_gather_u64offset_s64(...) __builtin_sve_svld1sw_gather_u64offset_s64(__VA_ARGS__) -#define svld1sw_vnum_u64(...) __builtin_sve_svld1sw_vnum_u64(__VA_ARGS__) -#define svld1sw_vnum_s64(...) __builtin_sve_svld1sw_vnum_s64(__VA_ARGS__) -#define svld1sw_u64(...) __builtin_sve_svld1sw_u64(__VA_ARGS__) -#define svld1sw_s64(...) __builtin_sve_svld1sw_s64(__VA_ARGS__) -#define svld1ub_gather_u32base_offset_u32(...) __builtin_sve_svld1ub_gather_u32base_offset_u32(__VA_ARGS__) -#define svld1ub_gather_u64base_offset_u64(...) __builtin_sve_svld1ub_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1ub_gather_u32base_offset_s32(...) __builtin_sve_svld1ub_gather_u32base_offset_s32(__VA_ARGS__) -#define svld1ub_gather_u64base_offset_s64(...) __builtin_sve_svld1ub_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1ub_gather_u32base_u32(...) __builtin_sve_svld1ub_gather_u32base_u32(__VA_ARGS__) -#define svld1ub_gather_u64base_u64(...) __builtin_sve_svld1ub_gather_u64base_u64(__VA_ARGS__) -#define svld1ub_gather_u32base_s32(...) __builtin_sve_svld1ub_gather_u32base_s32(__VA_ARGS__) -#define svld1ub_gather_u64base_s64(...) __builtin_sve_svld1ub_gather_u64base_s64(__VA_ARGS__) -#define svld1ub_gather_s32offset_u32(...) __builtin_sve_svld1ub_gather_s32offset_u32(__VA_ARGS__) -#define svld1ub_gather_s32offset_s32(...) __builtin_sve_svld1ub_gather_s32offset_s32(__VA_ARGS__) -#define svld1ub_gather_u32offset_u32(...) __builtin_sve_svld1ub_gather_u32offset_u32(__VA_ARGS__) -#define svld1ub_gather_u32offset_s32(...) __builtin_sve_svld1ub_gather_u32offset_s32(__VA_ARGS__) -#define svld1ub_gather_s64offset_u64(...) __builtin_sve_svld1ub_gather_s64offset_u64(__VA_ARGS__) -#define svld1ub_gather_s64offset_s64(...) __builtin_sve_svld1ub_gather_s64offset_s64(__VA_ARGS__) -#define svld1ub_gather_u64offset_u64(...) __builtin_sve_svld1ub_gather_u64offset_u64(__VA_ARGS__) -#define svld1ub_gather_u64offset_s64(...) __builtin_sve_svld1ub_gather_u64offset_s64(__VA_ARGS__) -#define svld1ub_vnum_u32(...) __builtin_sve_svld1ub_vnum_u32(__VA_ARGS__) -#define svld1ub_vnum_u64(...) __builtin_sve_svld1ub_vnum_u64(__VA_ARGS__) -#define svld1ub_vnum_u16(...) __builtin_sve_svld1ub_vnum_u16(__VA_ARGS__) -#define svld1ub_vnum_s32(...) __builtin_sve_svld1ub_vnum_s32(__VA_ARGS__) -#define svld1ub_vnum_s64(...) __builtin_sve_svld1ub_vnum_s64(__VA_ARGS__) -#define svld1ub_vnum_s16(...) __builtin_sve_svld1ub_vnum_s16(__VA_ARGS__) -#define svld1ub_u32(...) __builtin_sve_svld1ub_u32(__VA_ARGS__) -#define svld1ub_u64(...) __builtin_sve_svld1ub_u64(__VA_ARGS__) -#define svld1ub_u16(...) __builtin_sve_svld1ub_u16(__VA_ARGS__) -#define svld1ub_s32(...) __builtin_sve_svld1ub_s32(__VA_ARGS__) -#define svld1ub_s64(...) __builtin_sve_svld1ub_s64(__VA_ARGS__) -#define svld1ub_s16(...) __builtin_sve_svld1ub_s16(__VA_ARGS__) -#define svld1uh_gather_u32base_index_u32(...) __builtin_sve_svld1uh_gather_u32base_index_u32(__VA_ARGS__) -#define svld1uh_gather_u64base_index_u64(...) __builtin_sve_svld1uh_gather_u64base_index_u64(__VA_ARGS__) -#define svld1uh_gather_u32base_index_s32(...) __builtin_sve_svld1uh_gather_u32base_index_s32(__VA_ARGS__) -#define svld1uh_gather_u64base_index_s64(...) __builtin_sve_svld1uh_gather_u64base_index_s64(__VA_ARGS__) -#define svld1uh_gather_u32base_offset_u32(...) __builtin_sve_svld1uh_gather_u32base_offset_u32(__VA_ARGS__) -#define svld1uh_gather_u64base_offset_u64(...) __builtin_sve_svld1uh_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1uh_gather_u32base_offset_s32(...) __builtin_sve_svld1uh_gather_u32base_offset_s32(__VA_ARGS__) -#define svld1uh_gather_u64base_offset_s64(...) __builtin_sve_svld1uh_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1uh_gather_u32base_u32(...) __builtin_sve_svld1uh_gather_u32base_u32(__VA_ARGS__) -#define svld1uh_gather_u64base_u64(...) __builtin_sve_svld1uh_gather_u64base_u64(__VA_ARGS__) -#define svld1uh_gather_u32base_s32(...) __builtin_sve_svld1uh_gather_u32base_s32(__VA_ARGS__) -#define svld1uh_gather_u64base_s64(...) __builtin_sve_svld1uh_gather_u64base_s64(__VA_ARGS__) -#define svld1uh_gather_s32index_u32(...) __builtin_sve_svld1uh_gather_s32index_u32(__VA_ARGS__) -#define svld1uh_gather_s32index_s32(...) __builtin_sve_svld1uh_gather_s32index_s32(__VA_ARGS__) -#define svld1uh_gather_u32index_u32(...) __builtin_sve_svld1uh_gather_u32index_u32(__VA_ARGS__) -#define svld1uh_gather_u32index_s32(...) __builtin_sve_svld1uh_gather_u32index_s32(__VA_ARGS__) -#define svld1uh_gather_s64index_u64(...) __builtin_sve_svld1uh_gather_s64index_u64(__VA_ARGS__) -#define svld1uh_gather_s64index_s64(...) __builtin_sve_svld1uh_gather_s64index_s64(__VA_ARGS__) -#define svld1uh_gather_u64index_u64(...) __builtin_sve_svld1uh_gather_u64index_u64(__VA_ARGS__) -#define svld1uh_gather_u64index_s64(...) __builtin_sve_svld1uh_gather_u64index_s64(__VA_ARGS__) -#define svld1uh_gather_s32offset_u32(...) __builtin_sve_svld1uh_gather_s32offset_u32(__VA_ARGS__) -#define svld1uh_gather_s32offset_s32(...) __builtin_sve_svld1uh_gather_s32offset_s32(__VA_ARGS__) -#define svld1uh_gather_u32offset_u32(...) __builtin_sve_svld1uh_gather_u32offset_u32(__VA_ARGS__) -#define svld1uh_gather_u32offset_s32(...) __builtin_sve_svld1uh_gather_u32offset_s32(__VA_ARGS__) -#define svld1uh_gather_s64offset_u64(...) __builtin_sve_svld1uh_gather_s64offset_u64(__VA_ARGS__) -#define svld1uh_gather_s64offset_s64(...) __builtin_sve_svld1uh_gather_s64offset_s64(__VA_ARGS__) -#define svld1uh_gather_u64offset_u64(...) __builtin_sve_svld1uh_gather_u64offset_u64(__VA_ARGS__) -#define svld1uh_gather_u64offset_s64(...) __builtin_sve_svld1uh_gather_u64offset_s64(__VA_ARGS__) -#define svld1uh_vnum_u32(...) __builtin_sve_svld1uh_vnum_u32(__VA_ARGS__) -#define svld1uh_vnum_u64(...) __builtin_sve_svld1uh_vnum_u64(__VA_ARGS__) -#define svld1uh_vnum_s32(...) __builtin_sve_svld1uh_vnum_s32(__VA_ARGS__) -#define svld1uh_vnum_s64(...) __builtin_sve_svld1uh_vnum_s64(__VA_ARGS__) -#define svld1uh_u32(...) __builtin_sve_svld1uh_u32(__VA_ARGS__) -#define svld1uh_u64(...) __builtin_sve_svld1uh_u64(__VA_ARGS__) -#define svld1uh_s32(...) __builtin_sve_svld1uh_s32(__VA_ARGS__) -#define svld1uh_s64(...) __builtin_sve_svld1uh_s64(__VA_ARGS__) -#define svld1uw_gather_u64base_index_u64(...) __builtin_sve_svld1uw_gather_u64base_index_u64(__VA_ARGS__) -#define svld1uw_gather_u64base_index_s64(...) __builtin_sve_svld1uw_gather_u64base_index_s64(__VA_ARGS__) -#define svld1uw_gather_u64base_offset_u64(...) __builtin_sve_svld1uw_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1uw_gather_u64base_offset_s64(...) __builtin_sve_svld1uw_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1uw_gather_u64base_u64(...) __builtin_sve_svld1uw_gather_u64base_u64(__VA_ARGS__) -#define svld1uw_gather_u64base_s64(...) __builtin_sve_svld1uw_gather_u64base_s64(__VA_ARGS__) -#define svld1uw_gather_s64index_u64(...) __builtin_sve_svld1uw_gather_s64index_u64(__VA_ARGS__) -#define svld1uw_gather_s64index_s64(...) __builtin_sve_svld1uw_gather_s64index_s64(__VA_ARGS__) -#define svld1uw_gather_u64index_u64(...) __builtin_sve_svld1uw_gather_u64index_u64(__VA_ARGS__) -#define svld1uw_gather_u64index_s64(...) __builtin_sve_svld1uw_gather_u64index_s64(__VA_ARGS__) -#define svld1uw_gather_s64offset_u64(...) __builtin_sve_svld1uw_gather_s64offset_u64(__VA_ARGS__) -#define svld1uw_gather_s64offset_s64(...) __builtin_sve_svld1uw_gather_s64offset_s64(__VA_ARGS__) -#define svld1uw_gather_u64offset_u64(...) __builtin_sve_svld1uw_gather_u64offset_u64(__VA_ARGS__) -#define svld1uw_gather_u64offset_s64(...) __builtin_sve_svld1uw_gather_u64offset_s64(__VA_ARGS__) -#define svld1uw_vnum_u64(...) __builtin_sve_svld1uw_vnum_u64(__VA_ARGS__) -#define svld1uw_vnum_s64(...) __builtin_sve_svld1uw_vnum_s64(__VA_ARGS__) -#define svld1uw_u64(...) __builtin_sve_svld1uw_u64(__VA_ARGS__) -#define svld1uw_s64(...) __builtin_sve_svld1uw_s64(__VA_ARGS__) -#define svld2_u8(...) __builtin_sve_svld2_u8(__VA_ARGS__) -#define svld2_u32(...) __builtin_sve_svld2_u32(__VA_ARGS__) -#define svld2_u64(...) __builtin_sve_svld2_u64(__VA_ARGS__) -#define svld2_u16(...) __builtin_sve_svld2_u16(__VA_ARGS__) -#define svld2_s8(...) __builtin_sve_svld2_s8(__VA_ARGS__) -#define svld2_f64(...) __builtin_sve_svld2_f64(__VA_ARGS__) -#define svld2_f32(...) __builtin_sve_svld2_f32(__VA_ARGS__) -#define svld2_f16(...) __builtin_sve_svld2_f16(__VA_ARGS__) -#define svld2_s32(...) __builtin_sve_svld2_s32(__VA_ARGS__) -#define svld2_s64(...) __builtin_sve_svld2_s64(__VA_ARGS__) -#define svld2_s16(...) __builtin_sve_svld2_s16(__VA_ARGS__) -#define svld2_vnum_u8(...) __builtin_sve_svld2_vnum_u8(__VA_ARGS__) -#define svld2_vnum_u32(...) __builtin_sve_svld2_vnum_u32(__VA_ARGS__) -#define svld2_vnum_u64(...) __builtin_sve_svld2_vnum_u64(__VA_ARGS__) -#define svld2_vnum_u16(...) __builtin_sve_svld2_vnum_u16(__VA_ARGS__) -#define svld2_vnum_s8(...) __builtin_sve_svld2_vnum_s8(__VA_ARGS__) -#define svld2_vnum_f64(...) __builtin_sve_svld2_vnum_f64(__VA_ARGS__) -#define svld2_vnum_f32(...) __builtin_sve_svld2_vnum_f32(__VA_ARGS__) -#define svld2_vnum_f16(...) __builtin_sve_svld2_vnum_f16(__VA_ARGS__) -#define svld2_vnum_s32(...) __builtin_sve_svld2_vnum_s32(__VA_ARGS__) -#define svld2_vnum_s64(...) __builtin_sve_svld2_vnum_s64(__VA_ARGS__) -#define svld2_vnum_s16(...) __builtin_sve_svld2_vnum_s16(__VA_ARGS__) -#define svld3_u8(...) __builtin_sve_svld3_u8(__VA_ARGS__) -#define svld3_u32(...) __builtin_sve_svld3_u32(__VA_ARGS__) -#define svld3_u64(...) __builtin_sve_svld3_u64(__VA_ARGS__) -#define svld3_u16(...) __builtin_sve_svld3_u16(__VA_ARGS__) -#define svld3_s8(...) __builtin_sve_svld3_s8(__VA_ARGS__) -#define svld3_f64(...) __builtin_sve_svld3_f64(__VA_ARGS__) -#define svld3_f32(...) __builtin_sve_svld3_f32(__VA_ARGS__) -#define svld3_f16(...) __builtin_sve_svld3_f16(__VA_ARGS__) -#define svld3_s32(...) __builtin_sve_svld3_s32(__VA_ARGS__) -#define svld3_s64(...) __builtin_sve_svld3_s64(__VA_ARGS__) -#define svld3_s16(...) __builtin_sve_svld3_s16(__VA_ARGS__) -#define svld3_vnum_u8(...) __builtin_sve_svld3_vnum_u8(__VA_ARGS__) -#define svld3_vnum_u32(...) __builtin_sve_svld3_vnum_u32(__VA_ARGS__) -#define svld3_vnum_u64(...) __builtin_sve_svld3_vnum_u64(__VA_ARGS__) -#define svld3_vnum_u16(...) __builtin_sve_svld3_vnum_u16(__VA_ARGS__) -#define svld3_vnum_s8(...) __builtin_sve_svld3_vnum_s8(__VA_ARGS__) -#define svld3_vnum_f64(...) __builtin_sve_svld3_vnum_f64(__VA_ARGS__) -#define svld3_vnum_f32(...) __builtin_sve_svld3_vnum_f32(__VA_ARGS__) -#define svld3_vnum_f16(...) __builtin_sve_svld3_vnum_f16(__VA_ARGS__) -#define svld3_vnum_s32(...) __builtin_sve_svld3_vnum_s32(__VA_ARGS__) -#define svld3_vnum_s64(...) __builtin_sve_svld3_vnum_s64(__VA_ARGS__) -#define svld3_vnum_s16(...) __builtin_sve_svld3_vnum_s16(__VA_ARGS__) -#define svld4_u8(...) __builtin_sve_svld4_u8(__VA_ARGS__) -#define svld4_u32(...) __builtin_sve_svld4_u32(__VA_ARGS__) -#define svld4_u64(...) __builtin_sve_svld4_u64(__VA_ARGS__) -#define svld4_u16(...) __builtin_sve_svld4_u16(__VA_ARGS__) -#define svld4_s8(...) __builtin_sve_svld4_s8(__VA_ARGS__) -#define svld4_f64(...) __builtin_sve_svld4_f64(__VA_ARGS__) -#define svld4_f32(...) __builtin_sve_svld4_f32(__VA_ARGS__) -#define svld4_f16(...) __builtin_sve_svld4_f16(__VA_ARGS__) -#define svld4_s32(...) __builtin_sve_svld4_s32(__VA_ARGS__) -#define svld4_s64(...) __builtin_sve_svld4_s64(__VA_ARGS__) -#define svld4_s16(...) __builtin_sve_svld4_s16(__VA_ARGS__) -#define svld4_vnum_u8(...) __builtin_sve_svld4_vnum_u8(__VA_ARGS__) -#define svld4_vnum_u32(...) __builtin_sve_svld4_vnum_u32(__VA_ARGS__) -#define svld4_vnum_u64(...) __builtin_sve_svld4_vnum_u64(__VA_ARGS__) -#define svld4_vnum_u16(...) __builtin_sve_svld4_vnum_u16(__VA_ARGS__) -#define svld4_vnum_s8(...) __builtin_sve_svld4_vnum_s8(__VA_ARGS__) -#define svld4_vnum_f64(...) __builtin_sve_svld4_vnum_f64(__VA_ARGS__) -#define svld4_vnum_f32(...) __builtin_sve_svld4_vnum_f32(__VA_ARGS__) -#define svld4_vnum_f16(...) __builtin_sve_svld4_vnum_f16(__VA_ARGS__) -#define svld4_vnum_s32(...) __builtin_sve_svld4_vnum_s32(__VA_ARGS__) -#define svld4_vnum_s64(...) __builtin_sve_svld4_vnum_s64(__VA_ARGS__) -#define svld4_vnum_s16(...) __builtin_sve_svld4_vnum_s16(__VA_ARGS__) -#define svldff1_u8(...) __builtin_sve_svldff1_u8(__VA_ARGS__) -#define svldff1_u32(...) __builtin_sve_svldff1_u32(__VA_ARGS__) -#define svldff1_u64(...) __builtin_sve_svldff1_u64(__VA_ARGS__) -#define svldff1_u16(...) __builtin_sve_svldff1_u16(__VA_ARGS__) -#define svldff1_s8(...) __builtin_sve_svldff1_s8(__VA_ARGS__) -#define svldff1_f64(...) __builtin_sve_svldff1_f64(__VA_ARGS__) -#define svldff1_f32(...) __builtin_sve_svldff1_f32(__VA_ARGS__) -#define svldff1_f16(...) __builtin_sve_svldff1_f16(__VA_ARGS__) -#define svldff1_s32(...) __builtin_sve_svldff1_s32(__VA_ARGS__) -#define svldff1_s64(...) __builtin_sve_svldff1_s64(__VA_ARGS__) -#define svldff1_s16(...) __builtin_sve_svldff1_s16(__VA_ARGS__) -#define svldff1_gather_u32base_index_u32(...) __builtin_sve_svldff1_gather_u32base_index_u32(__VA_ARGS__) -#define svldff1_gather_u64base_index_u64(...) __builtin_sve_svldff1_gather_u64base_index_u64(__VA_ARGS__) -#define svldff1_gather_u64base_index_f64(...) __builtin_sve_svldff1_gather_u64base_index_f64(__VA_ARGS__) -#define svldff1_gather_u32base_index_f32(...) __builtin_sve_svldff1_gather_u32base_index_f32(__VA_ARGS__) -#define svldff1_gather_u32base_index_s32(...) __builtin_sve_svldff1_gather_u32base_index_s32(__VA_ARGS__) -#define svldff1_gather_u64base_index_s64(...) __builtin_sve_svldff1_gather_u64base_index_s64(__VA_ARGS__) -#define svldff1_gather_u32base_offset_u32(...) __builtin_sve_svldff1_gather_u32base_offset_u32(__VA_ARGS__) -#define svldff1_gather_u64base_offset_u64(...) __builtin_sve_svldff1_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1_gather_u64base_offset_f64(...) __builtin_sve_svldff1_gather_u64base_offset_f64(__VA_ARGS__) -#define svldff1_gather_u32base_offset_f32(...) __builtin_sve_svldff1_gather_u32base_offset_f32(__VA_ARGS__) -#define svldff1_gather_u32base_offset_s32(...) __builtin_sve_svldff1_gather_u32base_offset_s32(__VA_ARGS__) -#define svldff1_gather_u64base_offset_s64(...) __builtin_sve_svldff1_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1_gather_u32base_u32(...) __builtin_sve_svldff1_gather_u32base_u32(__VA_ARGS__) -#define svldff1_gather_u64base_u64(...) __builtin_sve_svldff1_gather_u64base_u64(__VA_ARGS__) -#define svldff1_gather_u64base_f64(...) __builtin_sve_svldff1_gather_u64base_f64(__VA_ARGS__) -#define svldff1_gather_u32base_f32(...) __builtin_sve_svldff1_gather_u32base_f32(__VA_ARGS__) -#define svldff1_gather_u32base_s32(...) __builtin_sve_svldff1_gather_u32base_s32(__VA_ARGS__) -#define svldff1_gather_u64base_s64(...) __builtin_sve_svldff1_gather_u64base_s64(__VA_ARGS__) -#define svldff1_gather_s32index_u32(...) __builtin_sve_svldff1_gather_s32index_u32(__VA_ARGS__) -#define svldff1_gather_s32index_f32(...) __builtin_sve_svldff1_gather_s32index_f32(__VA_ARGS__) -#define svldff1_gather_s32index_s32(...) __builtin_sve_svldff1_gather_s32index_s32(__VA_ARGS__) -#define svldff1_gather_u32index_u32(...) __builtin_sve_svldff1_gather_u32index_u32(__VA_ARGS__) -#define svldff1_gather_u32index_f32(...) __builtin_sve_svldff1_gather_u32index_f32(__VA_ARGS__) -#define svldff1_gather_u32index_s32(...) __builtin_sve_svldff1_gather_u32index_s32(__VA_ARGS__) -#define svldff1_gather_s64index_u64(...) __builtin_sve_svldff1_gather_s64index_u64(__VA_ARGS__) -#define svldff1_gather_s64index_f64(...) __builtin_sve_svldff1_gather_s64index_f64(__VA_ARGS__) -#define svldff1_gather_s64index_s64(...) __builtin_sve_svldff1_gather_s64index_s64(__VA_ARGS__) -#define svldff1_gather_u64index_u64(...) __builtin_sve_svldff1_gather_u64index_u64(__VA_ARGS__) -#define svldff1_gather_u64index_f64(...) __builtin_sve_svldff1_gather_u64index_f64(__VA_ARGS__) -#define svldff1_gather_u64index_s64(...) __builtin_sve_svldff1_gather_u64index_s64(__VA_ARGS__) -#define svldff1_gather_s32offset_u32(...) __builtin_sve_svldff1_gather_s32offset_u32(__VA_ARGS__) -#define svldff1_gather_s32offset_f32(...) __builtin_sve_svldff1_gather_s32offset_f32(__VA_ARGS__) -#define svldff1_gather_s32offset_s32(...) __builtin_sve_svldff1_gather_s32offset_s32(__VA_ARGS__) -#define svldff1_gather_u32offset_u32(...) __builtin_sve_svldff1_gather_u32offset_u32(__VA_ARGS__) -#define svldff1_gather_u32offset_f32(...) __builtin_sve_svldff1_gather_u32offset_f32(__VA_ARGS__) -#define svldff1_gather_u32offset_s32(...) __builtin_sve_svldff1_gather_u32offset_s32(__VA_ARGS__) -#define svldff1_gather_s64offset_u64(...) __builtin_sve_svldff1_gather_s64offset_u64(__VA_ARGS__) -#define svldff1_gather_s64offset_f64(...) __builtin_sve_svldff1_gather_s64offset_f64(__VA_ARGS__) -#define svldff1_gather_s64offset_s64(...) __builtin_sve_svldff1_gather_s64offset_s64(__VA_ARGS__) -#define svldff1_gather_u64offset_u64(...) __builtin_sve_svldff1_gather_u64offset_u64(__VA_ARGS__) -#define svldff1_gather_u64offset_f64(...) __builtin_sve_svldff1_gather_u64offset_f64(__VA_ARGS__) -#define svldff1_gather_u64offset_s64(...) __builtin_sve_svldff1_gather_u64offset_s64(__VA_ARGS__) -#define svldff1_vnum_u8(...) __builtin_sve_svldff1_vnum_u8(__VA_ARGS__) -#define svldff1_vnum_u32(...) __builtin_sve_svldff1_vnum_u32(__VA_ARGS__) -#define svldff1_vnum_u64(...) __builtin_sve_svldff1_vnum_u64(__VA_ARGS__) -#define svldff1_vnum_u16(...) __builtin_sve_svldff1_vnum_u16(__VA_ARGS__) -#define svldff1_vnum_s8(...) __builtin_sve_svldff1_vnum_s8(__VA_ARGS__) -#define svldff1_vnum_f64(...) __builtin_sve_svldff1_vnum_f64(__VA_ARGS__) -#define svldff1_vnum_f32(...) __builtin_sve_svldff1_vnum_f32(__VA_ARGS__) -#define svldff1_vnum_f16(...) __builtin_sve_svldff1_vnum_f16(__VA_ARGS__) -#define svldff1_vnum_s32(...) __builtin_sve_svldff1_vnum_s32(__VA_ARGS__) -#define svldff1_vnum_s64(...) __builtin_sve_svldff1_vnum_s64(__VA_ARGS__) -#define svldff1_vnum_s16(...) __builtin_sve_svldff1_vnum_s16(__VA_ARGS__) -#define svldff1sb_gather_u32base_offset_u32(...) __builtin_sve_svldff1sb_gather_u32base_offset_u32(__VA_ARGS__) -#define svldff1sb_gather_u64base_offset_u64(...) __builtin_sve_svldff1sb_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1sb_gather_u32base_offset_s32(...) __builtin_sve_svldff1sb_gather_u32base_offset_s32(__VA_ARGS__) -#define svldff1sb_gather_u64base_offset_s64(...) __builtin_sve_svldff1sb_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1sb_gather_u32base_u32(...) __builtin_sve_svldff1sb_gather_u32base_u32(__VA_ARGS__) -#define svldff1sb_gather_u64base_u64(...) __builtin_sve_svldff1sb_gather_u64base_u64(__VA_ARGS__) -#define svldff1sb_gather_u32base_s32(...) __builtin_sve_svldff1sb_gather_u32base_s32(__VA_ARGS__) -#define svldff1sb_gather_u64base_s64(...) __builtin_sve_svldff1sb_gather_u64base_s64(__VA_ARGS__) -#define svldff1sb_gather_s32offset_u32(...) __builtin_sve_svldff1sb_gather_s32offset_u32(__VA_ARGS__) -#define svldff1sb_gather_s32offset_s32(...) __builtin_sve_svldff1sb_gather_s32offset_s32(__VA_ARGS__) -#define svldff1sb_gather_u32offset_u32(...) __builtin_sve_svldff1sb_gather_u32offset_u32(__VA_ARGS__) -#define svldff1sb_gather_u32offset_s32(...) __builtin_sve_svldff1sb_gather_u32offset_s32(__VA_ARGS__) -#define svldff1sb_gather_s64offset_u64(...) __builtin_sve_svldff1sb_gather_s64offset_u64(__VA_ARGS__) -#define svldff1sb_gather_s64offset_s64(...) __builtin_sve_svldff1sb_gather_s64offset_s64(__VA_ARGS__) -#define svldff1sb_gather_u64offset_u64(...) __builtin_sve_svldff1sb_gather_u64offset_u64(__VA_ARGS__) -#define svldff1sb_gather_u64offset_s64(...) __builtin_sve_svldff1sb_gather_u64offset_s64(__VA_ARGS__) -#define svldff1sb_vnum_u32(...) __builtin_sve_svldff1sb_vnum_u32(__VA_ARGS__) -#define svldff1sb_vnum_u64(...) __builtin_sve_svldff1sb_vnum_u64(__VA_ARGS__) -#define svldff1sb_vnum_u16(...) __builtin_sve_svldff1sb_vnum_u16(__VA_ARGS__) -#define svldff1sb_vnum_s32(...) __builtin_sve_svldff1sb_vnum_s32(__VA_ARGS__) -#define svldff1sb_vnum_s64(...) __builtin_sve_svldff1sb_vnum_s64(__VA_ARGS__) -#define svldff1sb_vnum_s16(...) __builtin_sve_svldff1sb_vnum_s16(__VA_ARGS__) -#define svldff1sb_u32(...) __builtin_sve_svldff1sb_u32(__VA_ARGS__) -#define svldff1sb_u64(...) __builtin_sve_svldff1sb_u64(__VA_ARGS__) -#define svldff1sb_u16(...) __builtin_sve_svldff1sb_u16(__VA_ARGS__) -#define svldff1sb_s32(...) __builtin_sve_svldff1sb_s32(__VA_ARGS__) -#define svldff1sb_s64(...) __builtin_sve_svldff1sb_s64(__VA_ARGS__) -#define svldff1sb_s16(...) __builtin_sve_svldff1sb_s16(__VA_ARGS__) -#define svldff1sh_gather_u32base_index_u32(...) __builtin_sve_svldff1sh_gather_u32base_index_u32(__VA_ARGS__) -#define svldff1sh_gather_u64base_index_u64(...) __builtin_sve_svldff1sh_gather_u64base_index_u64(__VA_ARGS__) -#define svldff1sh_gather_u32base_index_s32(...) __builtin_sve_svldff1sh_gather_u32base_index_s32(__VA_ARGS__) -#define svldff1sh_gather_u64base_index_s64(...) __builtin_sve_svldff1sh_gather_u64base_index_s64(__VA_ARGS__) -#define svldff1sh_gather_u32base_offset_u32(...) __builtin_sve_svldff1sh_gather_u32base_offset_u32(__VA_ARGS__) -#define svldff1sh_gather_u64base_offset_u64(...) __builtin_sve_svldff1sh_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1sh_gather_u32base_offset_s32(...) __builtin_sve_svldff1sh_gather_u32base_offset_s32(__VA_ARGS__) -#define svldff1sh_gather_u64base_offset_s64(...) __builtin_sve_svldff1sh_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1sh_gather_u32base_u32(...) __builtin_sve_svldff1sh_gather_u32base_u32(__VA_ARGS__) -#define svldff1sh_gather_u64base_u64(...) __builtin_sve_svldff1sh_gather_u64base_u64(__VA_ARGS__) -#define svldff1sh_gather_u32base_s32(...) __builtin_sve_svldff1sh_gather_u32base_s32(__VA_ARGS__) -#define svldff1sh_gather_u64base_s64(...) __builtin_sve_svldff1sh_gather_u64base_s64(__VA_ARGS__) -#define svldff1sh_gather_s32index_u32(...) __builtin_sve_svldff1sh_gather_s32index_u32(__VA_ARGS__) -#define svldff1sh_gather_s32index_s32(...) __builtin_sve_svldff1sh_gather_s32index_s32(__VA_ARGS__) -#define svldff1sh_gather_u32index_u32(...) __builtin_sve_svldff1sh_gather_u32index_u32(__VA_ARGS__) -#define svldff1sh_gather_u32index_s32(...) __builtin_sve_svldff1sh_gather_u32index_s32(__VA_ARGS__) -#define svldff1sh_gather_s64index_u64(...) __builtin_sve_svldff1sh_gather_s64index_u64(__VA_ARGS__) -#define svldff1sh_gather_s64index_s64(...) __builtin_sve_svldff1sh_gather_s64index_s64(__VA_ARGS__) -#define svldff1sh_gather_u64index_u64(...) __builtin_sve_svldff1sh_gather_u64index_u64(__VA_ARGS__) -#define svldff1sh_gather_u64index_s64(...) __builtin_sve_svldff1sh_gather_u64index_s64(__VA_ARGS__) -#define svldff1sh_gather_s32offset_u32(...) __builtin_sve_svldff1sh_gather_s32offset_u32(__VA_ARGS__) -#define svldff1sh_gather_s32offset_s32(...) __builtin_sve_svldff1sh_gather_s32offset_s32(__VA_ARGS__) -#define svldff1sh_gather_u32offset_u32(...) __builtin_sve_svldff1sh_gather_u32offset_u32(__VA_ARGS__) -#define svldff1sh_gather_u32offset_s32(...) __builtin_sve_svldff1sh_gather_u32offset_s32(__VA_ARGS__) -#define svldff1sh_gather_s64offset_u64(...) __builtin_sve_svldff1sh_gather_s64offset_u64(__VA_ARGS__) -#define svldff1sh_gather_s64offset_s64(...) __builtin_sve_svldff1sh_gather_s64offset_s64(__VA_ARGS__) -#define svldff1sh_gather_u64offset_u64(...) __builtin_sve_svldff1sh_gather_u64offset_u64(__VA_ARGS__) -#define svldff1sh_gather_u64offset_s64(...) __builtin_sve_svldff1sh_gather_u64offset_s64(__VA_ARGS__) -#define svldff1sh_vnum_u32(...) __builtin_sve_svldff1sh_vnum_u32(__VA_ARGS__) -#define svldff1sh_vnum_u64(...) __builtin_sve_svldff1sh_vnum_u64(__VA_ARGS__) -#define svldff1sh_vnum_s32(...) __builtin_sve_svldff1sh_vnum_s32(__VA_ARGS__) -#define svldff1sh_vnum_s64(...) __builtin_sve_svldff1sh_vnum_s64(__VA_ARGS__) -#define svldff1sh_u32(...) __builtin_sve_svldff1sh_u32(__VA_ARGS__) -#define svldff1sh_u64(...) __builtin_sve_svldff1sh_u64(__VA_ARGS__) -#define svldff1sh_s32(...) __builtin_sve_svldff1sh_s32(__VA_ARGS__) -#define svldff1sh_s64(...) __builtin_sve_svldff1sh_s64(__VA_ARGS__) -#define svldff1sw_gather_u64base_index_u64(...) __builtin_sve_svldff1sw_gather_u64base_index_u64(__VA_ARGS__) -#define svldff1sw_gather_u64base_index_s64(...) __builtin_sve_svldff1sw_gather_u64base_index_s64(__VA_ARGS__) -#define svldff1sw_gather_u64base_offset_u64(...) __builtin_sve_svldff1sw_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1sw_gather_u64base_offset_s64(...) __builtin_sve_svldff1sw_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1sw_gather_u64base_u64(...) __builtin_sve_svldff1sw_gather_u64base_u64(__VA_ARGS__) -#define svldff1sw_gather_u64base_s64(...) __builtin_sve_svldff1sw_gather_u64base_s64(__VA_ARGS__) -#define svldff1sw_gather_s64index_u64(...) __builtin_sve_svldff1sw_gather_s64index_u64(__VA_ARGS__) -#define svldff1sw_gather_s64index_s64(...) __builtin_sve_svldff1sw_gather_s64index_s64(__VA_ARGS__) -#define svldff1sw_gather_u64index_u64(...) __builtin_sve_svldff1sw_gather_u64index_u64(__VA_ARGS__) -#define svldff1sw_gather_u64index_s64(...) __builtin_sve_svldff1sw_gather_u64index_s64(__VA_ARGS__) -#define svldff1sw_gather_s64offset_u64(...) __builtin_sve_svldff1sw_gather_s64offset_u64(__VA_ARGS__) -#define svldff1sw_gather_s64offset_s64(...) __builtin_sve_svldff1sw_gather_s64offset_s64(__VA_ARGS__) -#define svldff1sw_gather_u64offset_u64(...) __builtin_sve_svldff1sw_gather_u64offset_u64(__VA_ARGS__) -#define svldff1sw_gather_u64offset_s64(...) __builtin_sve_svldff1sw_gather_u64offset_s64(__VA_ARGS__) -#define svldff1sw_vnum_u64(...) __builtin_sve_svldff1sw_vnum_u64(__VA_ARGS__) -#define svldff1sw_vnum_s64(...) __builtin_sve_svldff1sw_vnum_s64(__VA_ARGS__) -#define svldff1sw_u64(...) __builtin_sve_svldff1sw_u64(__VA_ARGS__) -#define svldff1sw_s64(...) __builtin_sve_svldff1sw_s64(__VA_ARGS__) -#define svldff1ub_gather_u32base_offset_u32(...) __builtin_sve_svldff1ub_gather_u32base_offset_u32(__VA_ARGS__) -#define svldff1ub_gather_u64base_offset_u64(...) __builtin_sve_svldff1ub_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1ub_gather_u32base_offset_s32(...) __builtin_sve_svldff1ub_gather_u32base_offset_s32(__VA_ARGS__) -#define svldff1ub_gather_u64base_offset_s64(...) __builtin_sve_svldff1ub_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1ub_gather_u32base_u32(...) __builtin_sve_svldff1ub_gather_u32base_u32(__VA_ARGS__) -#define svldff1ub_gather_u64base_u64(...) __builtin_sve_svldff1ub_gather_u64base_u64(__VA_ARGS__) -#define svldff1ub_gather_u32base_s32(...) __builtin_sve_svldff1ub_gather_u32base_s32(__VA_ARGS__) -#define svldff1ub_gather_u64base_s64(...) __builtin_sve_svldff1ub_gather_u64base_s64(__VA_ARGS__) -#define svldff1ub_gather_s32offset_u32(...) __builtin_sve_svldff1ub_gather_s32offset_u32(__VA_ARGS__) -#define svldff1ub_gather_s32offset_s32(...) __builtin_sve_svldff1ub_gather_s32offset_s32(__VA_ARGS__) -#define svldff1ub_gather_u32offset_u32(...) __builtin_sve_svldff1ub_gather_u32offset_u32(__VA_ARGS__) -#define svldff1ub_gather_u32offset_s32(...) __builtin_sve_svldff1ub_gather_u32offset_s32(__VA_ARGS__) -#define svldff1ub_gather_s64offset_u64(...) __builtin_sve_svldff1ub_gather_s64offset_u64(__VA_ARGS__) -#define svldff1ub_gather_s64offset_s64(...) __builtin_sve_svldff1ub_gather_s64offset_s64(__VA_ARGS__) -#define svldff1ub_gather_u64offset_u64(...) __builtin_sve_svldff1ub_gather_u64offset_u64(__VA_ARGS__) -#define svldff1ub_gather_u64offset_s64(...) __builtin_sve_svldff1ub_gather_u64offset_s64(__VA_ARGS__) -#define svldff1ub_vnum_u32(...) __builtin_sve_svldff1ub_vnum_u32(__VA_ARGS__) -#define svldff1ub_vnum_u64(...) __builtin_sve_svldff1ub_vnum_u64(__VA_ARGS__) -#define svldff1ub_vnum_u16(...) __builtin_sve_svldff1ub_vnum_u16(__VA_ARGS__) -#define svldff1ub_vnum_s32(...) __builtin_sve_svldff1ub_vnum_s32(__VA_ARGS__) -#define svldff1ub_vnum_s64(...) __builtin_sve_svldff1ub_vnum_s64(__VA_ARGS__) -#define svldff1ub_vnum_s16(...) __builtin_sve_svldff1ub_vnum_s16(__VA_ARGS__) -#define svldff1ub_u32(...) __builtin_sve_svldff1ub_u32(__VA_ARGS__) -#define svldff1ub_u64(...) __builtin_sve_svldff1ub_u64(__VA_ARGS__) -#define svldff1ub_u16(...) __builtin_sve_svldff1ub_u16(__VA_ARGS__) -#define svldff1ub_s32(...) __builtin_sve_svldff1ub_s32(__VA_ARGS__) -#define svldff1ub_s64(...) __builtin_sve_svldff1ub_s64(__VA_ARGS__) -#define svldff1ub_s16(...) __builtin_sve_svldff1ub_s16(__VA_ARGS__) -#define svldff1uh_gather_u32base_index_u32(...) __builtin_sve_svldff1uh_gather_u32base_index_u32(__VA_ARGS__) -#define svldff1uh_gather_u64base_index_u64(...) __builtin_sve_svldff1uh_gather_u64base_index_u64(__VA_ARGS__) -#define svldff1uh_gather_u32base_index_s32(...) __builtin_sve_svldff1uh_gather_u32base_index_s32(__VA_ARGS__) -#define svldff1uh_gather_u64base_index_s64(...) __builtin_sve_svldff1uh_gather_u64base_index_s64(__VA_ARGS__) -#define svldff1uh_gather_u32base_offset_u32(...) __builtin_sve_svldff1uh_gather_u32base_offset_u32(__VA_ARGS__) -#define svldff1uh_gather_u64base_offset_u64(...) __builtin_sve_svldff1uh_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1uh_gather_u32base_offset_s32(...) __builtin_sve_svldff1uh_gather_u32base_offset_s32(__VA_ARGS__) -#define svldff1uh_gather_u64base_offset_s64(...) __builtin_sve_svldff1uh_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1uh_gather_u32base_u32(...) __builtin_sve_svldff1uh_gather_u32base_u32(__VA_ARGS__) -#define svldff1uh_gather_u64base_u64(...) __builtin_sve_svldff1uh_gather_u64base_u64(__VA_ARGS__) -#define svldff1uh_gather_u32base_s32(...) __builtin_sve_svldff1uh_gather_u32base_s32(__VA_ARGS__) -#define svldff1uh_gather_u64base_s64(...) __builtin_sve_svldff1uh_gather_u64base_s64(__VA_ARGS__) -#define svldff1uh_gather_s32index_u32(...) __builtin_sve_svldff1uh_gather_s32index_u32(__VA_ARGS__) -#define svldff1uh_gather_s32index_s32(...) __builtin_sve_svldff1uh_gather_s32index_s32(__VA_ARGS__) -#define svldff1uh_gather_u32index_u32(...) __builtin_sve_svldff1uh_gather_u32index_u32(__VA_ARGS__) -#define svldff1uh_gather_u32index_s32(...) __builtin_sve_svldff1uh_gather_u32index_s32(__VA_ARGS__) -#define svldff1uh_gather_s64index_u64(...) __builtin_sve_svldff1uh_gather_s64index_u64(__VA_ARGS__) -#define svldff1uh_gather_s64index_s64(...) __builtin_sve_svldff1uh_gather_s64index_s64(__VA_ARGS__) -#define svldff1uh_gather_u64index_u64(...) __builtin_sve_svldff1uh_gather_u64index_u64(__VA_ARGS__) -#define svldff1uh_gather_u64index_s64(...) __builtin_sve_svldff1uh_gather_u64index_s64(__VA_ARGS__) -#define svldff1uh_gather_s32offset_u32(...) __builtin_sve_svldff1uh_gather_s32offset_u32(__VA_ARGS__) -#define svldff1uh_gather_s32offset_s32(...) __builtin_sve_svldff1uh_gather_s32offset_s32(__VA_ARGS__) -#define svldff1uh_gather_u32offset_u32(...) __builtin_sve_svldff1uh_gather_u32offset_u32(__VA_ARGS__) -#define svldff1uh_gather_u32offset_s32(...) __builtin_sve_svldff1uh_gather_u32offset_s32(__VA_ARGS__) -#define svldff1uh_gather_s64offset_u64(...) __builtin_sve_svldff1uh_gather_s64offset_u64(__VA_ARGS__) -#define svldff1uh_gather_s64offset_s64(...) __builtin_sve_svldff1uh_gather_s64offset_s64(__VA_ARGS__) -#define svldff1uh_gather_u64offset_u64(...) __builtin_sve_svldff1uh_gather_u64offset_u64(__VA_ARGS__) -#define svldff1uh_gather_u64offset_s64(...) __builtin_sve_svldff1uh_gather_u64offset_s64(__VA_ARGS__) -#define svldff1uh_vnum_u32(...) __builtin_sve_svldff1uh_vnum_u32(__VA_ARGS__) -#define svldff1uh_vnum_u64(...) __builtin_sve_svldff1uh_vnum_u64(__VA_ARGS__) -#define svldff1uh_vnum_s32(...) __builtin_sve_svldff1uh_vnum_s32(__VA_ARGS__) -#define svldff1uh_vnum_s64(...) __builtin_sve_svldff1uh_vnum_s64(__VA_ARGS__) -#define svldff1uh_u32(...) __builtin_sve_svldff1uh_u32(__VA_ARGS__) -#define svldff1uh_u64(...) __builtin_sve_svldff1uh_u64(__VA_ARGS__) -#define svldff1uh_s32(...) __builtin_sve_svldff1uh_s32(__VA_ARGS__) -#define svldff1uh_s64(...) __builtin_sve_svldff1uh_s64(__VA_ARGS__) -#define svldff1uw_gather_u64base_index_u64(...) __builtin_sve_svldff1uw_gather_u64base_index_u64(__VA_ARGS__) -#define svldff1uw_gather_u64base_index_s64(...) __builtin_sve_svldff1uw_gather_u64base_index_s64(__VA_ARGS__) -#define svldff1uw_gather_u64base_offset_u64(...) __builtin_sve_svldff1uw_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1uw_gather_u64base_offset_s64(...) __builtin_sve_svldff1uw_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1uw_gather_u64base_u64(...) __builtin_sve_svldff1uw_gather_u64base_u64(__VA_ARGS__) -#define svldff1uw_gather_u64base_s64(...) __builtin_sve_svldff1uw_gather_u64base_s64(__VA_ARGS__) -#define svldff1uw_gather_s64index_u64(...) __builtin_sve_svldff1uw_gather_s64index_u64(__VA_ARGS__) -#define svldff1uw_gather_s64index_s64(...) __builtin_sve_svldff1uw_gather_s64index_s64(__VA_ARGS__) -#define svldff1uw_gather_u64index_u64(...) __builtin_sve_svldff1uw_gather_u64index_u64(__VA_ARGS__) -#define svldff1uw_gather_u64index_s64(...) __builtin_sve_svldff1uw_gather_u64index_s64(__VA_ARGS__) -#define svldff1uw_gather_s64offset_u64(...) __builtin_sve_svldff1uw_gather_s64offset_u64(__VA_ARGS__) -#define svldff1uw_gather_s64offset_s64(...) __builtin_sve_svldff1uw_gather_s64offset_s64(__VA_ARGS__) -#define svldff1uw_gather_u64offset_u64(...) __builtin_sve_svldff1uw_gather_u64offset_u64(__VA_ARGS__) -#define svldff1uw_gather_u64offset_s64(...) __builtin_sve_svldff1uw_gather_u64offset_s64(__VA_ARGS__) -#define svldff1uw_vnum_u64(...) __builtin_sve_svldff1uw_vnum_u64(__VA_ARGS__) -#define svldff1uw_vnum_s64(...) __builtin_sve_svldff1uw_vnum_s64(__VA_ARGS__) -#define svldff1uw_u64(...) __builtin_sve_svldff1uw_u64(__VA_ARGS__) -#define svldff1uw_s64(...) __builtin_sve_svldff1uw_s64(__VA_ARGS__) -#define svldnf1_u8(...) __builtin_sve_svldnf1_u8(__VA_ARGS__) -#define svldnf1_u32(...) __builtin_sve_svldnf1_u32(__VA_ARGS__) -#define svldnf1_u64(...) __builtin_sve_svldnf1_u64(__VA_ARGS__) -#define svldnf1_u16(...) __builtin_sve_svldnf1_u16(__VA_ARGS__) -#define svldnf1_s8(...) __builtin_sve_svldnf1_s8(__VA_ARGS__) -#define svldnf1_f64(...) __builtin_sve_svldnf1_f64(__VA_ARGS__) -#define svldnf1_f32(...) __builtin_sve_svldnf1_f32(__VA_ARGS__) -#define svldnf1_f16(...) __builtin_sve_svldnf1_f16(__VA_ARGS__) -#define svldnf1_s32(...) __builtin_sve_svldnf1_s32(__VA_ARGS__) -#define svldnf1_s64(...) __builtin_sve_svldnf1_s64(__VA_ARGS__) -#define svldnf1_s16(...) __builtin_sve_svldnf1_s16(__VA_ARGS__) -#define svldnf1_vnum_u8(...) __builtin_sve_svldnf1_vnum_u8(__VA_ARGS__) -#define svldnf1_vnum_u32(...) __builtin_sve_svldnf1_vnum_u32(__VA_ARGS__) -#define svldnf1_vnum_u64(...) __builtin_sve_svldnf1_vnum_u64(__VA_ARGS__) -#define svldnf1_vnum_u16(...) __builtin_sve_svldnf1_vnum_u16(__VA_ARGS__) -#define svldnf1_vnum_s8(...) __builtin_sve_svldnf1_vnum_s8(__VA_ARGS__) -#define svldnf1_vnum_f64(...) __builtin_sve_svldnf1_vnum_f64(__VA_ARGS__) -#define svldnf1_vnum_f32(...) __builtin_sve_svldnf1_vnum_f32(__VA_ARGS__) -#define svldnf1_vnum_f16(...) __builtin_sve_svldnf1_vnum_f16(__VA_ARGS__) -#define svldnf1_vnum_s32(...) __builtin_sve_svldnf1_vnum_s32(__VA_ARGS__) -#define svldnf1_vnum_s64(...) __builtin_sve_svldnf1_vnum_s64(__VA_ARGS__) -#define svldnf1_vnum_s16(...) __builtin_sve_svldnf1_vnum_s16(__VA_ARGS__) -#define svldnf1sb_vnum_u32(...) __builtin_sve_svldnf1sb_vnum_u32(__VA_ARGS__) -#define svldnf1sb_vnum_u64(...) __builtin_sve_svldnf1sb_vnum_u64(__VA_ARGS__) -#define svldnf1sb_vnum_u16(...) __builtin_sve_svldnf1sb_vnum_u16(__VA_ARGS__) -#define svldnf1sb_vnum_s32(...) __builtin_sve_svldnf1sb_vnum_s32(__VA_ARGS__) -#define svldnf1sb_vnum_s64(...) __builtin_sve_svldnf1sb_vnum_s64(__VA_ARGS__) -#define svldnf1sb_vnum_s16(...) __builtin_sve_svldnf1sb_vnum_s16(__VA_ARGS__) -#define svldnf1sb_u32(...) __builtin_sve_svldnf1sb_u32(__VA_ARGS__) -#define svldnf1sb_u64(...) __builtin_sve_svldnf1sb_u64(__VA_ARGS__) -#define svldnf1sb_u16(...) __builtin_sve_svldnf1sb_u16(__VA_ARGS__) -#define svldnf1sb_s32(...) __builtin_sve_svldnf1sb_s32(__VA_ARGS__) -#define svldnf1sb_s64(...) __builtin_sve_svldnf1sb_s64(__VA_ARGS__) -#define svldnf1sb_s16(...) __builtin_sve_svldnf1sb_s16(__VA_ARGS__) -#define svldnf1sh_vnum_u32(...) __builtin_sve_svldnf1sh_vnum_u32(__VA_ARGS__) -#define svldnf1sh_vnum_u64(...) __builtin_sve_svldnf1sh_vnum_u64(__VA_ARGS__) -#define svldnf1sh_vnum_s32(...) __builtin_sve_svldnf1sh_vnum_s32(__VA_ARGS__) -#define svldnf1sh_vnum_s64(...) __builtin_sve_svldnf1sh_vnum_s64(__VA_ARGS__) -#define svldnf1sh_u32(...) __builtin_sve_svldnf1sh_u32(__VA_ARGS__) -#define svldnf1sh_u64(...) __builtin_sve_svldnf1sh_u64(__VA_ARGS__) -#define svldnf1sh_s32(...) __builtin_sve_svldnf1sh_s32(__VA_ARGS__) -#define svldnf1sh_s64(...) __builtin_sve_svldnf1sh_s64(__VA_ARGS__) -#define svldnf1sw_vnum_u64(...) __builtin_sve_svldnf1sw_vnum_u64(__VA_ARGS__) -#define svldnf1sw_vnum_s64(...) __builtin_sve_svldnf1sw_vnum_s64(__VA_ARGS__) -#define svldnf1sw_u64(...) __builtin_sve_svldnf1sw_u64(__VA_ARGS__) -#define svldnf1sw_s64(...) __builtin_sve_svldnf1sw_s64(__VA_ARGS__) -#define svldnf1ub_vnum_u32(...) __builtin_sve_svldnf1ub_vnum_u32(__VA_ARGS__) -#define svldnf1ub_vnum_u64(...) __builtin_sve_svldnf1ub_vnum_u64(__VA_ARGS__) -#define svldnf1ub_vnum_u16(...) __builtin_sve_svldnf1ub_vnum_u16(__VA_ARGS__) -#define svldnf1ub_vnum_s32(...) __builtin_sve_svldnf1ub_vnum_s32(__VA_ARGS__) -#define svldnf1ub_vnum_s64(...) __builtin_sve_svldnf1ub_vnum_s64(__VA_ARGS__) -#define svldnf1ub_vnum_s16(...) __builtin_sve_svldnf1ub_vnum_s16(__VA_ARGS__) -#define svldnf1ub_u32(...) __builtin_sve_svldnf1ub_u32(__VA_ARGS__) -#define svldnf1ub_u64(...) __builtin_sve_svldnf1ub_u64(__VA_ARGS__) -#define svldnf1ub_u16(...) __builtin_sve_svldnf1ub_u16(__VA_ARGS__) -#define svldnf1ub_s32(...) __builtin_sve_svldnf1ub_s32(__VA_ARGS__) -#define svldnf1ub_s64(...) __builtin_sve_svldnf1ub_s64(__VA_ARGS__) -#define svldnf1ub_s16(...) __builtin_sve_svldnf1ub_s16(__VA_ARGS__) -#define svldnf1uh_vnum_u32(...) __builtin_sve_svldnf1uh_vnum_u32(__VA_ARGS__) -#define svldnf1uh_vnum_u64(...) __builtin_sve_svldnf1uh_vnum_u64(__VA_ARGS__) -#define svldnf1uh_vnum_s32(...) __builtin_sve_svldnf1uh_vnum_s32(__VA_ARGS__) -#define svldnf1uh_vnum_s64(...) __builtin_sve_svldnf1uh_vnum_s64(__VA_ARGS__) -#define svldnf1uh_u32(...) __builtin_sve_svldnf1uh_u32(__VA_ARGS__) -#define svldnf1uh_u64(...) __builtin_sve_svldnf1uh_u64(__VA_ARGS__) -#define svldnf1uh_s32(...) __builtin_sve_svldnf1uh_s32(__VA_ARGS__) -#define svldnf1uh_s64(...) __builtin_sve_svldnf1uh_s64(__VA_ARGS__) -#define svldnf1uw_vnum_u64(...) __builtin_sve_svldnf1uw_vnum_u64(__VA_ARGS__) -#define svldnf1uw_vnum_s64(...) __builtin_sve_svldnf1uw_vnum_s64(__VA_ARGS__) -#define svldnf1uw_u64(...) __builtin_sve_svldnf1uw_u64(__VA_ARGS__) -#define svldnf1uw_s64(...) __builtin_sve_svldnf1uw_s64(__VA_ARGS__) -#define svldnt1_u8(...) __builtin_sve_svldnt1_u8(__VA_ARGS__) -#define svldnt1_u32(...) __builtin_sve_svldnt1_u32(__VA_ARGS__) -#define svldnt1_u64(...) __builtin_sve_svldnt1_u64(__VA_ARGS__) -#define svldnt1_u16(...) __builtin_sve_svldnt1_u16(__VA_ARGS__) -#define svldnt1_s8(...) __builtin_sve_svldnt1_s8(__VA_ARGS__) -#define svldnt1_f64(...) __builtin_sve_svldnt1_f64(__VA_ARGS__) -#define svldnt1_f32(...) __builtin_sve_svldnt1_f32(__VA_ARGS__) -#define svldnt1_f16(...) __builtin_sve_svldnt1_f16(__VA_ARGS__) -#define svldnt1_s32(...) __builtin_sve_svldnt1_s32(__VA_ARGS__) -#define svldnt1_s64(...) __builtin_sve_svldnt1_s64(__VA_ARGS__) -#define svldnt1_s16(...) __builtin_sve_svldnt1_s16(__VA_ARGS__) -#define svldnt1_vnum_u8(...) __builtin_sve_svldnt1_vnum_u8(__VA_ARGS__) -#define svldnt1_vnum_u32(...) __builtin_sve_svldnt1_vnum_u32(__VA_ARGS__) -#define svldnt1_vnum_u64(...) __builtin_sve_svldnt1_vnum_u64(__VA_ARGS__) -#define svldnt1_vnum_u16(...) __builtin_sve_svldnt1_vnum_u16(__VA_ARGS__) -#define svldnt1_vnum_s8(...) __builtin_sve_svldnt1_vnum_s8(__VA_ARGS__) -#define svldnt1_vnum_f64(...) __builtin_sve_svldnt1_vnum_f64(__VA_ARGS__) -#define svldnt1_vnum_f32(...) __builtin_sve_svldnt1_vnum_f32(__VA_ARGS__) -#define svldnt1_vnum_f16(...) __builtin_sve_svldnt1_vnum_f16(__VA_ARGS__) -#define svldnt1_vnum_s32(...) __builtin_sve_svldnt1_vnum_s32(__VA_ARGS__) -#define svldnt1_vnum_s64(...) __builtin_sve_svldnt1_vnum_s64(__VA_ARGS__) -#define svldnt1_vnum_s16(...) __builtin_sve_svldnt1_vnum_s16(__VA_ARGS__) -#define svlen_u8(...) __builtin_sve_svlen_u8(__VA_ARGS__) -#define svlen_u32(...) __builtin_sve_svlen_u32(__VA_ARGS__) -#define svlen_u64(...) __builtin_sve_svlen_u64(__VA_ARGS__) -#define svlen_u16(...) __builtin_sve_svlen_u16(__VA_ARGS__) -#define svlen_s8(...) __builtin_sve_svlen_s8(__VA_ARGS__) -#define svlen_f64(...) __builtin_sve_svlen_f64(__VA_ARGS__) -#define svlen_f32(...) __builtin_sve_svlen_f32(__VA_ARGS__) -#define svlen_f16(...) __builtin_sve_svlen_f16(__VA_ARGS__) -#define svlen_s32(...) __builtin_sve_svlen_s32(__VA_ARGS__) -#define svlen_s64(...) __builtin_sve_svlen_s64(__VA_ARGS__) -#define svlen_s16(...) __builtin_sve_svlen_s16(__VA_ARGS__) -#define svlsl_n_u8_m(...) __builtin_sve_svlsl_n_u8_m(__VA_ARGS__) -#define svlsl_n_u32_m(...) __builtin_sve_svlsl_n_u32_m(__VA_ARGS__) -#define svlsl_n_u64_m(...) __builtin_sve_svlsl_n_u64_m(__VA_ARGS__) -#define svlsl_n_u16_m(...) __builtin_sve_svlsl_n_u16_m(__VA_ARGS__) -#define svlsl_n_s8_m(...) __builtin_sve_svlsl_n_s8_m(__VA_ARGS__) -#define svlsl_n_s32_m(...) __builtin_sve_svlsl_n_s32_m(__VA_ARGS__) -#define svlsl_n_s64_m(...) __builtin_sve_svlsl_n_s64_m(__VA_ARGS__) -#define svlsl_n_s16_m(...) __builtin_sve_svlsl_n_s16_m(__VA_ARGS__) -#define svlsl_n_u8_x(...) __builtin_sve_svlsl_n_u8_x(__VA_ARGS__) -#define svlsl_n_u32_x(...) __builtin_sve_svlsl_n_u32_x(__VA_ARGS__) -#define svlsl_n_u64_x(...) __builtin_sve_svlsl_n_u64_x(__VA_ARGS__) -#define svlsl_n_u16_x(...) __builtin_sve_svlsl_n_u16_x(__VA_ARGS__) -#define svlsl_n_s8_x(...) __builtin_sve_svlsl_n_s8_x(__VA_ARGS__) -#define svlsl_n_s32_x(...) __builtin_sve_svlsl_n_s32_x(__VA_ARGS__) -#define svlsl_n_s64_x(...) __builtin_sve_svlsl_n_s64_x(__VA_ARGS__) -#define svlsl_n_s16_x(...) __builtin_sve_svlsl_n_s16_x(__VA_ARGS__) -#define svlsl_n_u8_z(...) __builtin_sve_svlsl_n_u8_z(__VA_ARGS__) -#define svlsl_n_u32_z(...) __builtin_sve_svlsl_n_u32_z(__VA_ARGS__) -#define svlsl_n_u64_z(...) __builtin_sve_svlsl_n_u64_z(__VA_ARGS__) -#define svlsl_n_u16_z(...) __builtin_sve_svlsl_n_u16_z(__VA_ARGS__) -#define svlsl_n_s8_z(...) __builtin_sve_svlsl_n_s8_z(__VA_ARGS__) -#define svlsl_n_s32_z(...) __builtin_sve_svlsl_n_s32_z(__VA_ARGS__) -#define svlsl_n_s64_z(...) __builtin_sve_svlsl_n_s64_z(__VA_ARGS__) -#define svlsl_n_s16_z(...) __builtin_sve_svlsl_n_s16_z(__VA_ARGS__) -#define svlsl_u8_m(...) __builtin_sve_svlsl_u8_m(__VA_ARGS__) -#define svlsl_u32_m(...) __builtin_sve_svlsl_u32_m(__VA_ARGS__) -#define svlsl_u64_m(...) __builtin_sve_svlsl_u64_m(__VA_ARGS__) -#define svlsl_u16_m(...) __builtin_sve_svlsl_u16_m(__VA_ARGS__) -#define svlsl_s8_m(...) __builtin_sve_svlsl_s8_m(__VA_ARGS__) -#define svlsl_s32_m(...) __builtin_sve_svlsl_s32_m(__VA_ARGS__) -#define svlsl_s64_m(...) __builtin_sve_svlsl_s64_m(__VA_ARGS__) -#define svlsl_s16_m(...) __builtin_sve_svlsl_s16_m(__VA_ARGS__) -#define svlsl_u8_x(...) __builtin_sve_svlsl_u8_x(__VA_ARGS__) -#define svlsl_u32_x(...) __builtin_sve_svlsl_u32_x(__VA_ARGS__) -#define svlsl_u64_x(...) __builtin_sve_svlsl_u64_x(__VA_ARGS__) -#define svlsl_u16_x(...) __builtin_sve_svlsl_u16_x(__VA_ARGS__) -#define svlsl_s8_x(...) __builtin_sve_svlsl_s8_x(__VA_ARGS__) -#define svlsl_s32_x(...) __builtin_sve_svlsl_s32_x(__VA_ARGS__) -#define svlsl_s64_x(...) __builtin_sve_svlsl_s64_x(__VA_ARGS__) -#define svlsl_s16_x(...) __builtin_sve_svlsl_s16_x(__VA_ARGS__) -#define svlsl_u8_z(...) __builtin_sve_svlsl_u8_z(__VA_ARGS__) -#define svlsl_u32_z(...) __builtin_sve_svlsl_u32_z(__VA_ARGS__) -#define svlsl_u64_z(...) __builtin_sve_svlsl_u64_z(__VA_ARGS__) -#define svlsl_u16_z(...) __builtin_sve_svlsl_u16_z(__VA_ARGS__) -#define svlsl_s8_z(...) __builtin_sve_svlsl_s8_z(__VA_ARGS__) -#define svlsl_s32_z(...) __builtin_sve_svlsl_s32_z(__VA_ARGS__) -#define svlsl_s64_z(...) __builtin_sve_svlsl_s64_z(__VA_ARGS__) -#define svlsl_s16_z(...) __builtin_sve_svlsl_s16_z(__VA_ARGS__) -#define svlsl_wide_n_u8_m(...) __builtin_sve_svlsl_wide_n_u8_m(__VA_ARGS__) -#define svlsl_wide_n_u32_m(...) __builtin_sve_svlsl_wide_n_u32_m(__VA_ARGS__) -#define svlsl_wide_n_u16_m(...) __builtin_sve_svlsl_wide_n_u16_m(__VA_ARGS__) -#define svlsl_wide_n_s8_m(...) __builtin_sve_svlsl_wide_n_s8_m(__VA_ARGS__) -#define svlsl_wide_n_s32_m(...) __builtin_sve_svlsl_wide_n_s32_m(__VA_ARGS__) -#define svlsl_wide_n_s16_m(...) __builtin_sve_svlsl_wide_n_s16_m(__VA_ARGS__) -#define svlsl_wide_n_u8_x(...) __builtin_sve_svlsl_wide_n_u8_x(__VA_ARGS__) -#define svlsl_wide_n_u32_x(...) __builtin_sve_svlsl_wide_n_u32_x(__VA_ARGS__) -#define svlsl_wide_n_u16_x(...) __builtin_sve_svlsl_wide_n_u16_x(__VA_ARGS__) -#define svlsl_wide_n_s8_x(...) __builtin_sve_svlsl_wide_n_s8_x(__VA_ARGS__) -#define svlsl_wide_n_s32_x(...) __builtin_sve_svlsl_wide_n_s32_x(__VA_ARGS__) -#define svlsl_wide_n_s16_x(...) __builtin_sve_svlsl_wide_n_s16_x(__VA_ARGS__) -#define svlsl_wide_n_u8_z(...) __builtin_sve_svlsl_wide_n_u8_z(__VA_ARGS__) -#define svlsl_wide_n_u32_z(...) __builtin_sve_svlsl_wide_n_u32_z(__VA_ARGS__) -#define svlsl_wide_n_u16_z(...) __builtin_sve_svlsl_wide_n_u16_z(__VA_ARGS__) -#define svlsl_wide_n_s8_z(...) __builtin_sve_svlsl_wide_n_s8_z(__VA_ARGS__) -#define svlsl_wide_n_s32_z(...) __builtin_sve_svlsl_wide_n_s32_z(__VA_ARGS__) -#define svlsl_wide_n_s16_z(...) __builtin_sve_svlsl_wide_n_s16_z(__VA_ARGS__) -#define svlsl_wide_u8_m(...) __builtin_sve_svlsl_wide_u8_m(__VA_ARGS__) -#define svlsl_wide_u32_m(...) __builtin_sve_svlsl_wide_u32_m(__VA_ARGS__) -#define svlsl_wide_u16_m(...) __builtin_sve_svlsl_wide_u16_m(__VA_ARGS__) -#define svlsl_wide_s8_m(...) __builtin_sve_svlsl_wide_s8_m(__VA_ARGS__) -#define svlsl_wide_s32_m(...) __builtin_sve_svlsl_wide_s32_m(__VA_ARGS__) -#define svlsl_wide_s16_m(...) __builtin_sve_svlsl_wide_s16_m(__VA_ARGS__) -#define svlsl_wide_u8_x(...) __builtin_sve_svlsl_wide_u8_x(__VA_ARGS__) -#define svlsl_wide_u32_x(...) __builtin_sve_svlsl_wide_u32_x(__VA_ARGS__) -#define svlsl_wide_u16_x(...) __builtin_sve_svlsl_wide_u16_x(__VA_ARGS__) -#define svlsl_wide_s8_x(...) __builtin_sve_svlsl_wide_s8_x(__VA_ARGS__) -#define svlsl_wide_s32_x(...) __builtin_sve_svlsl_wide_s32_x(__VA_ARGS__) -#define svlsl_wide_s16_x(...) __builtin_sve_svlsl_wide_s16_x(__VA_ARGS__) -#define svlsl_wide_u8_z(...) __builtin_sve_svlsl_wide_u8_z(__VA_ARGS__) -#define svlsl_wide_u32_z(...) __builtin_sve_svlsl_wide_u32_z(__VA_ARGS__) -#define svlsl_wide_u16_z(...) __builtin_sve_svlsl_wide_u16_z(__VA_ARGS__) -#define svlsl_wide_s8_z(...) __builtin_sve_svlsl_wide_s8_z(__VA_ARGS__) -#define svlsl_wide_s32_z(...) __builtin_sve_svlsl_wide_s32_z(__VA_ARGS__) -#define svlsl_wide_s16_z(...) __builtin_sve_svlsl_wide_s16_z(__VA_ARGS__) -#define svlsr_n_u8_m(...) __builtin_sve_svlsr_n_u8_m(__VA_ARGS__) -#define svlsr_n_u32_m(...) __builtin_sve_svlsr_n_u32_m(__VA_ARGS__) -#define svlsr_n_u64_m(...) __builtin_sve_svlsr_n_u64_m(__VA_ARGS__) -#define svlsr_n_u16_m(...) __builtin_sve_svlsr_n_u16_m(__VA_ARGS__) -#define svlsr_n_u8_x(...) __builtin_sve_svlsr_n_u8_x(__VA_ARGS__) -#define svlsr_n_u32_x(...) __builtin_sve_svlsr_n_u32_x(__VA_ARGS__) -#define svlsr_n_u64_x(...) __builtin_sve_svlsr_n_u64_x(__VA_ARGS__) -#define svlsr_n_u16_x(...) __builtin_sve_svlsr_n_u16_x(__VA_ARGS__) -#define svlsr_n_u8_z(...) __builtin_sve_svlsr_n_u8_z(__VA_ARGS__) -#define svlsr_n_u32_z(...) __builtin_sve_svlsr_n_u32_z(__VA_ARGS__) -#define svlsr_n_u64_z(...) __builtin_sve_svlsr_n_u64_z(__VA_ARGS__) -#define svlsr_n_u16_z(...) __builtin_sve_svlsr_n_u16_z(__VA_ARGS__) -#define svlsr_u8_m(...) __builtin_sve_svlsr_u8_m(__VA_ARGS__) -#define svlsr_u32_m(...) __builtin_sve_svlsr_u32_m(__VA_ARGS__) -#define svlsr_u64_m(...) __builtin_sve_svlsr_u64_m(__VA_ARGS__) -#define svlsr_u16_m(...) __builtin_sve_svlsr_u16_m(__VA_ARGS__) -#define svlsr_u8_x(...) __builtin_sve_svlsr_u8_x(__VA_ARGS__) -#define svlsr_u32_x(...) __builtin_sve_svlsr_u32_x(__VA_ARGS__) -#define svlsr_u64_x(...) __builtin_sve_svlsr_u64_x(__VA_ARGS__) -#define svlsr_u16_x(...) __builtin_sve_svlsr_u16_x(__VA_ARGS__) -#define svlsr_u8_z(...) __builtin_sve_svlsr_u8_z(__VA_ARGS__) -#define svlsr_u32_z(...) __builtin_sve_svlsr_u32_z(__VA_ARGS__) -#define svlsr_u64_z(...) __builtin_sve_svlsr_u64_z(__VA_ARGS__) -#define svlsr_u16_z(...) __builtin_sve_svlsr_u16_z(__VA_ARGS__) -#define svlsr_wide_n_u8_m(...) __builtin_sve_svlsr_wide_n_u8_m(__VA_ARGS__) -#define svlsr_wide_n_u32_m(...) __builtin_sve_svlsr_wide_n_u32_m(__VA_ARGS__) -#define svlsr_wide_n_u16_m(...) __builtin_sve_svlsr_wide_n_u16_m(__VA_ARGS__) -#define svlsr_wide_n_u8_x(...) __builtin_sve_svlsr_wide_n_u8_x(__VA_ARGS__) -#define svlsr_wide_n_u32_x(...) __builtin_sve_svlsr_wide_n_u32_x(__VA_ARGS__) -#define svlsr_wide_n_u16_x(...) __builtin_sve_svlsr_wide_n_u16_x(__VA_ARGS__) -#define svlsr_wide_n_u8_z(...) __builtin_sve_svlsr_wide_n_u8_z(__VA_ARGS__) -#define svlsr_wide_n_u32_z(...) __builtin_sve_svlsr_wide_n_u32_z(__VA_ARGS__) -#define svlsr_wide_n_u16_z(...) __builtin_sve_svlsr_wide_n_u16_z(__VA_ARGS__) -#define svlsr_wide_u8_m(...) __builtin_sve_svlsr_wide_u8_m(__VA_ARGS__) -#define svlsr_wide_u32_m(...) __builtin_sve_svlsr_wide_u32_m(__VA_ARGS__) -#define svlsr_wide_u16_m(...) __builtin_sve_svlsr_wide_u16_m(__VA_ARGS__) -#define svlsr_wide_u8_x(...) __builtin_sve_svlsr_wide_u8_x(__VA_ARGS__) -#define svlsr_wide_u32_x(...) __builtin_sve_svlsr_wide_u32_x(__VA_ARGS__) -#define svlsr_wide_u16_x(...) __builtin_sve_svlsr_wide_u16_x(__VA_ARGS__) -#define svlsr_wide_u8_z(...) __builtin_sve_svlsr_wide_u8_z(__VA_ARGS__) -#define svlsr_wide_u32_z(...) __builtin_sve_svlsr_wide_u32_z(__VA_ARGS__) -#define svlsr_wide_u16_z(...) __builtin_sve_svlsr_wide_u16_z(__VA_ARGS__) -#define svmad_n_f64_m(...) __builtin_sve_svmad_n_f64_m(__VA_ARGS__) -#define svmad_n_f32_m(...) __builtin_sve_svmad_n_f32_m(__VA_ARGS__) -#define svmad_n_f16_m(...) __builtin_sve_svmad_n_f16_m(__VA_ARGS__) -#define svmad_n_f64_x(...) __builtin_sve_svmad_n_f64_x(__VA_ARGS__) -#define svmad_n_f32_x(...) __builtin_sve_svmad_n_f32_x(__VA_ARGS__) -#define svmad_n_f16_x(...) __builtin_sve_svmad_n_f16_x(__VA_ARGS__) -#define svmad_n_f64_z(...) __builtin_sve_svmad_n_f64_z(__VA_ARGS__) -#define svmad_n_f32_z(...) __builtin_sve_svmad_n_f32_z(__VA_ARGS__) -#define svmad_n_f16_z(...) __builtin_sve_svmad_n_f16_z(__VA_ARGS__) -#define svmad_n_u8_m(...) __builtin_sve_svmad_n_u8_m(__VA_ARGS__) -#define svmad_n_u32_m(...) __builtin_sve_svmad_n_u32_m(__VA_ARGS__) -#define svmad_n_u64_m(...) __builtin_sve_svmad_n_u64_m(__VA_ARGS__) -#define svmad_n_u16_m(...) __builtin_sve_svmad_n_u16_m(__VA_ARGS__) -#define svmad_n_s8_m(...) __builtin_sve_svmad_n_s8_m(__VA_ARGS__) -#define svmad_n_s32_m(...) __builtin_sve_svmad_n_s32_m(__VA_ARGS__) -#define svmad_n_s64_m(...) __builtin_sve_svmad_n_s64_m(__VA_ARGS__) -#define svmad_n_s16_m(...) __builtin_sve_svmad_n_s16_m(__VA_ARGS__) -#define svmad_n_u8_x(...) __builtin_sve_svmad_n_u8_x(__VA_ARGS__) -#define svmad_n_u32_x(...) __builtin_sve_svmad_n_u32_x(__VA_ARGS__) -#define svmad_n_u64_x(...) __builtin_sve_svmad_n_u64_x(__VA_ARGS__) -#define svmad_n_u16_x(...) __builtin_sve_svmad_n_u16_x(__VA_ARGS__) -#define svmad_n_s8_x(...) __builtin_sve_svmad_n_s8_x(__VA_ARGS__) -#define svmad_n_s32_x(...) __builtin_sve_svmad_n_s32_x(__VA_ARGS__) -#define svmad_n_s64_x(...) __builtin_sve_svmad_n_s64_x(__VA_ARGS__) -#define svmad_n_s16_x(...) __builtin_sve_svmad_n_s16_x(__VA_ARGS__) -#define svmad_n_u8_z(...) __builtin_sve_svmad_n_u8_z(__VA_ARGS__) -#define svmad_n_u32_z(...) __builtin_sve_svmad_n_u32_z(__VA_ARGS__) -#define svmad_n_u64_z(...) __builtin_sve_svmad_n_u64_z(__VA_ARGS__) -#define svmad_n_u16_z(...) __builtin_sve_svmad_n_u16_z(__VA_ARGS__) -#define svmad_n_s8_z(...) __builtin_sve_svmad_n_s8_z(__VA_ARGS__) -#define svmad_n_s32_z(...) __builtin_sve_svmad_n_s32_z(__VA_ARGS__) -#define svmad_n_s64_z(...) __builtin_sve_svmad_n_s64_z(__VA_ARGS__) -#define svmad_n_s16_z(...) __builtin_sve_svmad_n_s16_z(__VA_ARGS__) -#define svmad_f64_m(...) __builtin_sve_svmad_f64_m(__VA_ARGS__) -#define svmad_f32_m(...) __builtin_sve_svmad_f32_m(__VA_ARGS__) -#define svmad_f16_m(...) __builtin_sve_svmad_f16_m(__VA_ARGS__) -#define svmad_f64_x(...) __builtin_sve_svmad_f64_x(__VA_ARGS__) -#define svmad_f32_x(...) __builtin_sve_svmad_f32_x(__VA_ARGS__) -#define svmad_f16_x(...) __builtin_sve_svmad_f16_x(__VA_ARGS__) -#define svmad_f64_z(...) __builtin_sve_svmad_f64_z(__VA_ARGS__) -#define svmad_f32_z(...) __builtin_sve_svmad_f32_z(__VA_ARGS__) -#define svmad_f16_z(...) __builtin_sve_svmad_f16_z(__VA_ARGS__) -#define svmad_u8_m(...) __builtin_sve_svmad_u8_m(__VA_ARGS__) -#define svmad_u32_m(...) __builtin_sve_svmad_u32_m(__VA_ARGS__) -#define svmad_u64_m(...) __builtin_sve_svmad_u64_m(__VA_ARGS__) -#define svmad_u16_m(...) __builtin_sve_svmad_u16_m(__VA_ARGS__) -#define svmad_s8_m(...) __builtin_sve_svmad_s8_m(__VA_ARGS__) -#define svmad_s32_m(...) __builtin_sve_svmad_s32_m(__VA_ARGS__) -#define svmad_s64_m(...) __builtin_sve_svmad_s64_m(__VA_ARGS__) -#define svmad_s16_m(...) __builtin_sve_svmad_s16_m(__VA_ARGS__) -#define svmad_u8_x(...) __builtin_sve_svmad_u8_x(__VA_ARGS__) -#define svmad_u32_x(...) __builtin_sve_svmad_u32_x(__VA_ARGS__) -#define svmad_u64_x(...) __builtin_sve_svmad_u64_x(__VA_ARGS__) -#define svmad_u16_x(...) __builtin_sve_svmad_u16_x(__VA_ARGS__) -#define svmad_s8_x(...) __builtin_sve_svmad_s8_x(__VA_ARGS__) -#define svmad_s32_x(...) __builtin_sve_svmad_s32_x(__VA_ARGS__) -#define svmad_s64_x(...) __builtin_sve_svmad_s64_x(__VA_ARGS__) -#define svmad_s16_x(...) __builtin_sve_svmad_s16_x(__VA_ARGS__) -#define svmad_u8_z(...) __builtin_sve_svmad_u8_z(__VA_ARGS__) -#define svmad_u32_z(...) __builtin_sve_svmad_u32_z(__VA_ARGS__) -#define svmad_u64_z(...) __builtin_sve_svmad_u64_z(__VA_ARGS__) -#define svmad_u16_z(...) __builtin_sve_svmad_u16_z(__VA_ARGS__) -#define svmad_s8_z(...) __builtin_sve_svmad_s8_z(__VA_ARGS__) -#define svmad_s32_z(...) __builtin_sve_svmad_s32_z(__VA_ARGS__) -#define svmad_s64_z(...) __builtin_sve_svmad_s64_z(__VA_ARGS__) -#define svmad_s16_z(...) __builtin_sve_svmad_s16_z(__VA_ARGS__) -#define svmax_n_f64_m(...) __builtin_sve_svmax_n_f64_m(__VA_ARGS__) -#define svmax_n_f32_m(...) __builtin_sve_svmax_n_f32_m(__VA_ARGS__) -#define svmax_n_f16_m(...) __builtin_sve_svmax_n_f16_m(__VA_ARGS__) -#define svmax_n_f64_x(...) __builtin_sve_svmax_n_f64_x(__VA_ARGS__) -#define svmax_n_f32_x(...) __builtin_sve_svmax_n_f32_x(__VA_ARGS__) -#define svmax_n_f16_x(...) __builtin_sve_svmax_n_f16_x(__VA_ARGS__) -#define svmax_n_f64_z(...) __builtin_sve_svmax_n_f64_z(__VA_ARGS__) -#define svmax_n_f32_z(...) __builtin_sve_svmax_n_f32_z(__VA_ARGS__) -#define svmax_n_f16_z(...) __builtin_sve_svmax_n_f16_z(__VA_ARGS__) -#define svmax_n_s8_m(...) __builtin_sve_svmax_n_s8_m(__VA_ARGS__) -#define svmax_n_s32_m(...) __builtin_sve_svmax_n_s32_m(__VA_ARGS__) -#define svmax_n_s64_m(...) __builtin_sve_svmax_n_s64_m(__VA_ARGS__) -#define svmax_n_s16_m(...) __builtin_sve_svmax_n_s16_m(__VA_ARGS__) -#define svmax_n_s8_x(...) __builtin_sve_svmax_n_s8_x(__VA_ARGS__) -#define svmax_n_s32_x(...) __builtin_sve_svmax_n_s32_x(__VA_ARGS__) -#define svmax_n_s64_x(...) __builtin_sve_svmax_n_s64_x(__VA_ARGS__) -#define svmax_n_s16_x(...) __builtin_sve_svmax_n_s16_x(__VA_ARGS__) -#define svmax_n_s8_z(...) __builtin_sve_svmax_n_s8_z(__VA_ARGS__) -#define svmax_n_s32_z(...) __builtin_sve_svmax_n_s32_z(__VA_ARGS__) -#define svmax_n_s64_z(...) __builtin_sve_svmax_n_s64_z(__VA_ARGS__) -#define svmax_n_s16_z(...) __builtin_sve_svmax_n_s16_z(__VA_ARGS__) -#define svmax_n_u8_m(...) __builtin_sve_svmax_n_u8_m(__VA_ARGS__) -#define svmax_n_u32_m(...) __builtin_sve_svmax_n_u32_m(__VA_ARGS__) -#define svmax_n_u64_m(...) __builtin_sve_svmax_n_u64_m(__VA_ARGS__) -#define svmax_n_u16_m(...) __builtin_sve_svmax_n_u16_m(__VA_ARGS__) -#define svmax_n_u8_x(...) __builtin_sve_svmax_n_u8_x(__VA_ARGS__) -#define svmax_n_u32_x(...) __builtin_sve_svmax_n_u32_x(__VA_ARGS__) -#define svmax_n_u64_x(...) __builtin_sve_svmax_n_u64_x(__VA_ARGS__) -#define svmax_n_u16_x(...) __builtin_sve_svmax_n_u16_x(__VA_ARGS__) -#define svmax_n_u8_z(...) __builtin_sve_svmax_n_u8_z(__VA_ARGS__) -#define svmax_n_u32_z(...) __builtin_sve_svmax_n_u32_z(__VA_ARGS__) -#define svmax_n_u64_z(...) __builtin_sve_svmax_n_u64_z(__VA_ARGS__) -#define svmax_n_u16_z(...) __builtin_sve_svmax_n_u16_z(__VA_ARGS__) -#define svmax_f64_m(...) __builtin_sve_svmax_f64_m(__VA_ARGS__) -#define svmax_f32_m(...) __builtin_sve_svmax_f32_m(__VA_ARGS__) -#define svmax_f16_m(...) __builtin_sve_svmax_f16_m(__VA_ARGS__) -#define svmax_f64_x(...) __builtin_sve_svmax_f64_x(__VA_ARGS__) -#define svmax_f32_x(...) __builtin_sve_svmax_f32_x(__VA_ARGS__) -#define svmax_f16_x(...) __builtin_sve_svmax_f16_x(__VA_ARGS__) -#define svmax_f64_z(...) __builtin_sve_svmax_f64_z(__VA_ARGS__) -#define svmax_f32_z(...) __builtin_sve_svmax_f32_z(__VA_ARGS__) -#define svmax_f16_z(...) __builtin_sve_svmax_f16_z(__VA_ARGS__) -#define svmax_s8_m(...) __builtin_sve_svmax_s8_m(__VA_ARGS__) -#define svmax_s32_m(...) __builtin_sve_svmax_s32_m(__VA_ARGS__) -#define svmax_s64_m(...) __builtin_sve_svmax_s64_m(__VA_ARGS__) -#define svmax_s16_m(...) __builtin_sve_svmax_s16_m(__VA_ARGS__) -#define svmax_s8_x(...) __builtin_sve_svmax_s8_x(__VA_ARGS__) -#define svmax_s32_x(...) __builtin_sve_svmax_s32_x(__VA_ARGS__) -#define svmax_s64_x(...) __builtin_sve_svmax_s64_x(__VA_ARGS__) -#define svmax_s16_x(...) __builtin_sve_svmax_s16_x(__VA_ARGS__) -#define svmax_s8_z(...) __builtin_sve_svmax_s8_z(__VA_ARGS__) -#define svmax_s32_z(...) __builtin_sve_svmax_s32_z(__VA_ARGS__) -#define svmax_s64_z(...) __builtin_sve_svmax_s64_z(__VA_ARGS__) -#define svmax_s16_z(...) __builtin_sve_svmax_s16_z(__VA_ARGS__) -#define svmax_u8_m(...) __builtin_sve_svmax_u8_m(__VA_ARGS__) -#define svmax_u32_m(...) __builtin_sve_svmax_u32_m(__VA_ARGS__) -#define svmax_u64_m(...) __builtin_sve_svmax_u64_m(__VA_ARGS__) -#define svmax_u16_m(...) __builtin_sve_svmax_u16_m(__VA_ARGS__) -#define svmax_u8_x(...) __builtin_sve_svmax_u8_x(__VA_ARGS__) -#define svmax_u32_x(...) __builtin_sve_svmax_u32_x(__VA_ARGS__) -#define svmax_u64_x(...) __builtin_sve_svmax_u64_x(__VA_ARGS__) -#define svmax_u16_x(...) __builtin_sve_svmax_u16_x(__VA_ARGS__) -#define svmax_u8_z(...) __builtin_sve_svmax_u8_z(__VA_ARGS__) -#define svmax_u32_z(...) __builtin_sve_svmax_u32_z(__VA_ARGS__) -#define svmax_u64_z(...) __builtin_sve_svmax_u64_z(__VA_ARGS__) -#define svmax_u16_z(...) __builtin_sve_svmax_u16_z(__VA_ARGS__) -#define svmaxnm_n_f64_m(...) __builtin_sve_svmaxnm_n_f64_m(__VA_ARGS__) -#define svmaxnm_n_f32_m(...) __builtin_sve_svmaxnm_n_f32_m(__VA_ARGS__) -#define svmaxnm_n_f16_m(...) __builtin_sve_svmaxnm_n_f16_m(__VA_ARGS__) -#define svmaxnm_n_f64_x(...) __builtin_sve_svmaxnm_n_f64_x(__VA_ARGS__) -#define svmaxnm_n_f32_x(...) __builtin_sve_svmaxnm_n_f32_x(__VA_ARGS__) -#define svmaxnm_n_f16_x(...) __builtin_sve_svmaxnm_n_f16_x(__VA_ARGS__) -#define svmaxnm_n_f64_z(...) __builtin_sve_svmaxnm_n_f64_z(__VA_ARGS__) -#define svmaxnm_n_f32_z(...) __builtin_sve_svmaxnm_n_f32_z(__VA_ARGS__) -#define svmaxnm_n_f16_z(...) __builtin_sve_svmaxnm_n_f16_z(__VA_ARGS__) -#define svmaxnm_f64_m(...) __builtin_sve_svmaxnm_f64_m(__VA_ARGS__) -#define svmaxnm_f32_m(...) __builtin_sve_svmaxnm_f32_m(__VA_ARGS__) -#define svmaxnm_f16_m(...) __builtin_sve_svmaxnm_f16_m(__VA_ARGS__) -#define svmaxnm_f64_x(...) __builtin_sve_svmaxnm_f64_x(__VA_ARGS__) -#define svmaxnm_f32_x(...) __builtin_sve_svmaxnm_f32_x(__VA_ARGS__) -#define svmaxnm_f16_x(...) __builtin_sve_svmaxnm_f16_x(__VA_ARGS__) -#define svmaxnm_f64_z(...) __builtin_sve_svmaxnm_f64_z(__VA_ARGS__) -#define svmaxnm_f32_z(...) __builtin_sve_svmaxnm_f32_z(__VA_ARGS__) -#define svmaxnm_f16_z(...) __builtin_sve_svmaxnm_f16_z(__VA_ARGS__) -#define svmaxnmv_f64(...) __builtin_sve_svmaxnmv_f64(__VA_ARGS__) -#define svmaxnmv_f32(...) __builtin_sve_svmaxnmv_f32(__VA_ARGS__) -#define svmaxnmv_f16(...) __builtin_sve_svmaxnmv_f16(__VA_ARGS__) -#define svmaxv_f64(...) __builtin_sve_svmaxv_f64(__VA_ARGS__) -#define svmaxv_f32(...) __builtin_sve_svmaxv_f32(__VA_ARGS__) -#define svmaxv_f16(...) __builtin_sve_svmaxv_f16(__VA_ARGS__) -#define svmaxv_s8(...) __builtin_sve_svmaxv_s8(__VA_ARGS__) -#define svmaxv_s32(...) __builtin_sve_svmaxv_s32(__VA_ARGS__) -#define svmaxv_s64(...) __builtin_sve_svmaxv_s64(__VA_ARGS__) -#define svmaxv_s16(...) __builtin_sve_svmaxv_s16(__VA_ARGS__) -#define svmaxv_u8(...) __builtin_sve_svmaxv_u8(__VA_ARGS__) -#define svmaxv_u32(...) __builtin_sve_svmaxv_u32(__VA_ARGS__) -#define svmaxv_u64(...) __builtin_sve_svmaxv_u64(__VA_ARGS__) -#define svmaxv_u16(...) __builtin_sve_svmaxv_u16(__VA_ARGS__) -#define svmin_n_f64_m(...) __builtin_sve_svmin_n_f64_m(__VA_ARGS__) -#define svmin_n_f32_m(...) __builtin_sve_svmin_n_f32_m(__VA_ARGS__) -#define svmin_n_f16_m(...) __builtin_sve_svmin_n_f16_m(__VA_ARGS__) -#define svmin_n_f64_x(...) __builtin_sve_svmin_n_f64_x(__VA_ARGS__) -#define svmin_n_f32_x(...) __builtin_sve_svmin_n_f32_x(__VA_ARGS__) -#define svmin_n_f16_x(...) __builtin_sve_svmin_n_f16_x(__VA_ARGS__) -#define svmin_n_f64_z(...) __builtin_sve_svmin_n_f64_z(__VA_ARGS__) -#define svmin_n_f32_z(...) __builtin_sve_svmin_n_f32_z(__VA_ARGS__) -#define svmin_n_f16_z(...) __builtin_sve_svmin_n_f16_z(__VA_ARGS__) -#define svmin_n_s8_m(...) __builtin_sve_svmin_n_s8_m(__VA_ARGS__) -#define svmin_n_s32_m(...) __builtin_sve_svmin_n_s32_m(__VA_ARGS__) -#define svmin_n_s64_m(...) __builtin_sve_svmin_n_s64_m(__VA_ARGS__) -#define svmin_n_s16_m(...) __builtin_sve_svmin_n_s16_m(__VA_ARGS__) -#define svmin_n_s8_x(...) __builtin_sve_svmin_n_s8_x(__VA_ARGS__) -#define svmin_n_s32_x(...) __builtin_sve_svmin_n_s32_x(__VA_ARGS__) -#define svmin_n_s64_x(...) __builtin_sve_svmin_n_s64_x(__VA_ARGS__) -#define svmin_n_s16_x(...) __builtin_sve_svmin_n_s16_x(__VA_ARGS__) -#define svmin_n_s8_z(...) __builtin_sve_svmin_n_s8_z(__VA_ARGS__) -#define svmin_n_s32_z(...) __builtin_sve_svmin_n_s32_z(__VA_ARGS__) -#define svmin_n_s64_z(...) __builtin_sve_svmin_n_s64_z(__VA_ARGS__) -#define svmin_n_s16_z(...) __builtin_sve_svmin_n_s16_z(__VA_ARGS__) -#define svmin_n_u8_m(...) __builtin_sve_svmin_n_u8_m(__VA_ARGS__) -#define svmin_n_u32_m(...) __builtin_sve_svmin_n_u32_m(__VA_ARGS__) -#define svmin_n_u64_m(...) __builtin_sve_svmin_n_u64_m(__VA_ARGS__) -#define svmin_n_u16_m(...) __builtin_sve_svmin_n_u16_m(__VA_ARGS__) -#define svmin_n_u8_x(...) __builtin_sve_svmin_n_u8_x(__VA_ARGS__) -#define svmin_n_u32_x(...) __builtin_sve_svmin_n_u32_x(__VA_ARGS__) -#define svmin_n_u64_x(...) __builtin_sve_svmin_n_u64_x(__VA_ARGS__) -#define svmin_n_u16_x(...) __builtin_sve_svmin_n_u16_x(__VA_ARGS__) -#define svmin_n_u8_z(...) __builtin_sve_svmin_n_u8_z(__VA_ARGS__) -#define svmin_n_u32_z(...) __builtin_sve_svmin_n_u32_z(__VA_ARGS__) -#define svmin_n_u64_z(...) __builtin_sve_svmin_n_u64_z(__VA_ARGS__) -#define svmin_n_u16_z(...) __builtin_sve_svmin_n_u16_z(__VA_ARGS__) -#define svmin_f64_m(...) __builtin_sve_svmin_f64_m(__VA_ARGS__) -#define svmin_f32_m(...) __builtin_sve_svmin_f32_m(__VA_ARGS__) -#define svmin_f16_m(...) __builtin_sve_svmin_f16_m(__VA_ARGS__) -#define svmin_f64_x(...) __builtin_sve_svmin_f64_x(__VA_ARGS__) -#define svmin_f32_x(...) __builtin_sve_svmin_f32_x(__VA_ARGS__) -#define svmin_f16_x(...) __builtin_sve_svmin_f16_x(__VA_ARGS__) -#define svmin_f64_z(...) __builtin_sve_svmin_f64_z(__VA_ARGS__) -#define svmin_f32_z(...) __builtin_sve_svmin_f32_z(__VA_ARGS__) -#define svmin_f16_z(...) __builtin_sve_svmin_f16_z(__VA_ARGS__) -#define svmin_s8_m(...) __builtin_sve_svmin_s8_m(__VA_ARGS__) -#define svmin_s32_m(...) __builtin_sve_svmin_s32_m(__VA_ARGS__) -#define svmin_s64_m(...) __builtin_sve_svmin_s64_m(__VA_ARGS__) -#define svmin_s16_m(...) __builtin_sve_svmin_s16_m(__VA_ARGS__) -#define svmin_s8_x(...) __builtin_sve_svmin_s8_x(__VA_ARGS__) -#define svmin_s32_x(...) __builtin_sve_svmin_s32_x(__VA_ARGS__) -#define svmin_s64_x(...) __builtin_sve_svmin_s64_x(__VA_ARGS__) -#define svmin_s16_x(...) __builtin_sve_svmin_s16_x(__VA_ARGS__) -#define svmin_s8_z(...) __builtin_sve_svmin_s8_z(__VA_ARGS__) -#define svmin_s32_z(...) __builtin_sve_svmin_s32_z(__VA_ARGS__) -#define svmin_s64_z(...) __builtin_sve_svmin_s64_z(__VA_ARGS__) -#define svmin_s16_z(...) __builtin_sve_svmin_s16_z(__VA_ARGS__) -#define svmin_u8_m(...) __builtin_sve_svmin_u8_m(__VA_ARGS__) -#define svmin_u32_m(...) __builtin_sve_svmin_u32_m(__VA_ARGS__) -#define svmin_u64_m(...) __builtin_sve_svmin_u64_m(__VA_ARGS__) -#define svmin_u16_m(...) __builtin_sve_svmin_u16_m(__VA_ARGS__) -#define svmin_u8_x(...) __builtin_sve_svmin_u8_x(__VA_ARGS__) -#define svmin_u32_x(...) __builtin_sve_svmin_u32_x(__VA_ARGS__) -#define svmin_u64_x(...) __builtin_sve_svmin_u64_x(__VA_ARGS__) -#define svmin_u16_x(...) __builtin_sve_svmin_u16_x(__VA_ARGS__) -#define svmin_u8_z(...) __builtin_sve_svmin_u8_z(__VA_ARGS__) -#define svmin_u32_z(...) __builtin_sve_svmin_u32_z(__VA_ARGS__) -#define svmin_u64_z(...) __builtin_sve_svmin_u64_z(__VA_ARGS__) -#define svmin_u16_z(...) __builtin_sve_svmin_u16_z(__VA_ARGS__) -#define svminnm_n_f64_m(...) __builtin_sve_svminnm_n_f64_m(__VA_ARGS__) -#define svminnm_n_f32_m(...) __builtin_sve_svminnm_n_f32_m(__VA_ARGS__) -#define svminnm_n_f16_m(...) __builtin_sve_svminnm_n_f16_m(__VA_ARGS__) -#define svminnm_n_f64_x(...) __builtin_sve_svminnm_n_f64_x(__VA_ARGS__) -#define svminnm_n_f32_x(...) __builtin_sve_svminnm_n_f32_x(__VA_ARGS__) -#define svminnm_n_f16_x(...) __builtin_sve_svminnm_n_f16_x(__VA_ARGS__) -#define svminnm_n_f64_z(...) __builtin_sve_svminnm_n_f64_z(__VA_ARGS__) -#define svminnm_n_f32_z(...) __builtin_sve_svminnm_n_f32_z(__VA_ARGS__) -#define svminnm_n_f16_z(...) __builtin_sve_svminnm_n_f16_z(__VA_ARGS__) -#define svminnm_f64_m(...) __builtin_sve_svminnm_f64_m(__VA_ARGS__) -#define svminnm_f32_m(...) __builtin_sve_svminnm_f32_m(__VA_ARGS__) -#define svminnm_f16_m(...) __builtin_sve_svminnm_f16_m(__VA_ARGS__) -#define svminnm_f64_x(...) __builtin_sve_svminnm_f64_x(__VA_ARGS__) -#define svminnm_f32_x(...) __builtin_sve_svminnm_f32_x(__VA_ARGS__) -#define svminnm_f16_x(...) __builtin_sve_svminnm_f16_x(__VA_ARGS__) -#define svminnm_f64_z(...) __builtin_sve_svminnm_f64_z(__VA_ARGS__) -#define svminnm_f32_z(...) __builtin_sve_svminnm_f32_z(__VA_ARGS__) -#define svminnm_f16_z(...) __builtin_sve_svminnm_f16_z(__VA_ARGS__) -#define svminnmv_f64(...) __builtin_sve_svminnmv_f64(__VA_ARGS__) -#define svminnmv_f32(...) __builtin_sve_svminnmv_f32(__VA_ARGS__) -#define svminnmv_f16(...) __builtin_sve_svminnmv_f16(__VA_ARGS__) -#define svminv_f64(...) __builtin_sve_svminv_f64(__VA_ARGS__) -#define svminv_f32(...) __builtin_sve_svminv_f32(__VA_ARGS__) -#define svminv_f16(...) __builtin_sve_svminv_f16(__VA_ARGS__) -#define svminv_s8(...) __builtin_sve_svminv_s8(__VA_ARGS__) -#define svminv_s32(...) __builtin_sve_svminv_s32(__VA_ARGS__) -#define svminv_s64(...) __builtin_sve_svminv_s64(__VA_ARGS__) -#define svminv_s16(...) __builtin_sve_svminv_s16(__VA_ARGS__) -#define svminv_u8(...) __builtin_sve_svminv_u8(__VA_ARGS__) -#define svminv_u32(...) __builtin_sve_svminv_u32(__VA_ARGS__) -#define svminv_u64(...) __builtin_sve_svminv_u64(__VA_ARGS__) -#define svminv_u16(...) __builtin_sve_svminv_u16(__VA_ARGS__) -#define svmla_n_f64_m(...) __builtin_sve_svmla_n_f64_m(__VA_ARGS__) -#define svmla_n_f32_m(...) __builtin_sve_svmla_n_f32_m(__VA_ARGS__) -#define svmla_n_f16_m(...) __builtin_sve_svmla_n_f16_m(__VA_ARGS__) -#define svmla_n_f64_x(...) __builtin_sve_svmla_n_f64_x(__VA_ARGS__) -#define svmla_n_f32_x(...) __builtin_sve_svmla_n_f32_x(__VA_ARGS__) -#define svmla_n_f16_x(...) __builtin_sve_svmla_n_f16_x(__VA_ARGS__) -#define svmla_n_f64_z(...) __builtin_sve_svmla_n_f64_z(__VA_ARGS__) -#define svmla_n_f32_z(...) __builtin_sve_svmla_n_f32_z(__VA_ARGS__) -#define svmla_n_f16_z(...) __builtin_sve_svmla_n_f16_z(__VA_ARGS__) -#define svmla_n_u8_m(...) __builtin_sve_svmla_n_u8_m(__VA_ARGS__) -#define svmla_n_u32_m(...) __builtin_sve_svmla_n_u32_m(__VA_ARGS__) -#define svmla_n_u64_m(...) __builtin_sve_svmla_n_u64_m(__VA_ARGS__) -#define svmla_n_u16_m(...) __builtin_sve_svmla_n_u16_m(__VA_ARGS__) -#define svmla_n_s8_m(...) __builtin_sve_svmla_n_s8_m(__VA_ARGS__) -#define svmla_n_s32_m(...) __builtin_sve_svmla_n_s32_m(__VA_ARGS__) -#define svmla_n_s64_m(...) __builtin_sve_svmla_n_s64_m(__VA_ARGS__) -#define svmla_n_s16_m(...) __builtin_sve_svmla_n_s16_m(__VA_ARGS__) -#define svmla_n_u8_x(...) __builtin_sve_svmla_n_u8_x(__VA_ARGS__) -#define svmla_n_u32_x(...) __builtin_sve_svmla_n_u32_x(__VA_ARGS__) -#define svmla_n_u64_x(...) __builtin_sve_svmla_n_u64_x(__VA_ARGS__) -#define svmla_n_u16_x(...) __builtin_sve_svmla_n_u16_x(__VA_ARGS__) -#define svmla_n_s8_x(...) __builtin_sve_svmla_n_s8_x(__VA_ARGS__) -#define svmla_n_s32_x(...) __builtin_sve_svmla_n_s32_x(__VA_ARGS__) -#define svmla_n_s64_x(...) __builtin_sve_svmla_n_s64_x(__VA_ARGS__) -#define svmla_n_s16_x(...) __builtin_sve_svmla_n_s16_x(__VA_ARGS__) -#define svmla_n_u8_z(...) __builtin_sve_svmla_n_u8_z(__VA_ARGS__) -#define svmla_n_u32_z(...) __builtin_sve_svmla_n_u32_z(__VA_ARGS__) -#define svmla_n_u64_z(...) __builtin_sve_svmla_n_u64_z(__VA_ARGS__) -#define svmla_n_u16_z(...) __builtin_sve_svmla_n_u16_z(__VA_ARGS__) -#define svmla_n_s8_z(...) __builtin_sve_svmla_n_s8_z(__VA_ARGS__) -#define svmla_n_s32_z(...) __builtin_sve_svmla_n_s32_z(__VA_ARGS__) -#define svmla_n_s64_z(...) __builtin_sve_svmla_n_s64_z(__VA_ARGS__) -#define svmla_n_s16_z(...) __builtin_sve_svmla_n_s16_z(__VA_ARGS__) -#define svmla_f64_m(...) __builtin_sve_svmla_f64_m(__VA_ARGS__) -#define svmla_f32_m(...) __builtin_sve_svmla_f32_m(__VA_ARGS__) -#define svmla_f16_m(...) __builtin_sve_svmla_f16_m(__VA_ARGS__) -#define svmla_f64_x(...) __builtin_sve_svmla_f64_x(__VA_ARGS__) -#define svmla_f32_x(...) __builtin_sve_svmla_f32_x(__VA_ARGS__) -#define svmla_f16_x(...) __builtin_sve_svmla_f16_x(__VA_ARGS__) -#define svmla_f64_z(...) __builtin_sve_svmla_f64_z(__VA_ARGS__) -#define svmla_f32_z(...) __builtin_sve_svmla_f32_z(__VA_ARGS__) -#define svmla_f16_z(...) __builtin_sve_svmla_f16_z(__VA_ARGS__) -#define svmla_u8_m(...) __builtin_sve_svmla_u8_m(__VA_ARGS__) -#define svmla_u32_m(...) __builtin_sve_svmla_u32_m(__VA_ARGS__) -#define svmla_u64_m(...) __builtin_sve_svmla_u64_m(__VA_ARGS__) -#define svmla_u16_m(...) __builtin_sve_svmla_u16_m(__VA_ARGS__) -#define svmla_s8_m(...) __builtin_sve_svmla_s8_m(__VA_ARGS__) -#define svmla_s32_m(...) __builtin_sve_svmla_s32_m(__VA_ARGS__) -#define svmla_s64_m(...) __builtin_sve_svmla_s64_m(__VA_ARGS__) -#define svmla_s16_m(...) __builtin_sve_svmla_s16_m(__VA_ARGS__) -#define svmla_u8_x(...) __builtin_sve_svmla_u8_x(__VA_ARGS__) -#define svmla_u32_x(...) __builtin_sve_svmla_u32_x(__VA_ARGS__) -#define svmla_u64_x(...) __builtin_sve_svmla_u64_x(__VA_ARGS__) -#define svmla_u16_x(...) __builtin_sve_svmla_u16_x(__VA_ARGS__) -#define svmla_s8_x(...) __builtin_sve_svmla_s8_x(__VA_ARGS__) -#define svmla_s32_x(...) __builtin_sve_svmla_s32_x(__VA_ARGS__) -#define svmla_s64_x(...) __builtin_sve_svmla_s64_x(__VA_ARGS__) -#define svmla_s16_x(...) __builtin_sve_svmla_s16_x(__VA_ARGS__) -#define svmla_u8_z(...) __builtin_sve_svmla_u8_z(__VA_ARGS__) -#define svmla_u32_z(...) __builtin_sve_svmla_u32_z(__VA_ARGS__) -#define svmla_u64_z(...) __builtin_sve_svmla_u64_z(__VA_ARGS__) -#define svmla_u16_z(...) __builtin_sve_svmla_u16_z(__VA_ARGS__) -#define svmla_s8_z(...) __builtin_sve_svmla_s8_z(__VA_ARGS__) -#define svmla_s32_z(...) __builtin_sve_svmla_s32_z(__VA_ARGS__) -#define svmla_s64_z(...) __builtin_sve_svmla_s64_z(__VA_ARGS__) -#define svmla_s16_z(...) __builtin_sve_svmla_s16_z(__VA_ARGS__) -#define svmla_lane_f64(...) __builtin_sve_svmla_lane_f64(__VA_ARGS__) -#define svmla_lane_f32(...) __builtin_sve_svmla_lane_f32(__VA_ARGS__) -#define svmla_lane_f16(...) __builtin_sve_svmla_lane_f16(__VA_ARGS__) -#define svmls_n_f64_m(...) __builtin_sve_svmls_n_f64_m(__VA_ARGS__) -#define svmls_n_f32_m(...) __builtin_sve_svmls_n_f32_m(__VA_ARGS__) -#define svmls_n_f16_m(...) __builtin_sve_svmls_n_f16_m(__VA_ARGS__) -#define svmls_n_f64_x(...) __builtin_sve_svmls_n_f64_x(__VA_ARGS__) -#define svmls_n_f32_x(...) __builtin_sve_svmls_n_f32_x(__VA_ARGS__) -#define svmls_n_f16_x(...) __builtin_sve_svmls_n_f16_x(__VA_ARGS__) -#define svmls_n_f64_z(...) __builtin_sve_svmls_n_f64_z(__VA_ARGS__) -#define svmls_n_f32_z(...) __builtin_sve_svmls_n_f32_z(__VA_ARGS__) -#define svmls_n_f16_z(...) __builtin_sve_svmls_n_f16_z(__VA_ARGS__) -#define svmls_n_u8_m(...) __builtin_sve_svmls_n_u8_m(__VA_ARGS__) -#define svmls_n_u32_m(...) __builtin_sve_svmls_n_u32_m(__VA_ARGS__) -#define svmls_n_u64_m(...) __builtin_sve_svmls_n_u64_m(__VA_ARGS__) -#define svmls_n_u16_m(...) __builtin_sve_svmls_n_u16_m(__VA_ARGS__) -#define svmls_n_s8_m(...) __builtin_sve_svmls_n_s8_m(__VA_ARGS__) -#define svmls_n_s32_m(...) __builtin_sve_svmls_n_s32_m(__VA_ARGS__) -#define svmls_n_s64_m(...) __builtin_sve_svmls_n_s64_m(__VA_ARGS__) -#define svmls_n_s16_m(...) __builtin_sve_svmls_n_s16_m(__VA_ARGS__) -#define svmls_n_u8_x(...) __builtin_sve_svmls_n_u8_x(__VA_ARGS__) -#define svmls_n_u32_x(...) __builtin_sve_svmls_n_u32_x(__VA_ARGS__) -#define svmls_n_u64_x(...) __builtin_sve_svmls_n_u64_x(__VA_ARGS__) -#define svmls_n_u16_x(...) __builtin_sve_svmls_n_u16_x(__VA_ARGS__) -#define svmls_n_s8_x(...) __builtin_sve_svmls_n_s8_x(__VA_ARGS__) -#define svmls_n_s32_x(...) __builtin_sve_svmls_n_s32_x(__VA_ARGS__) -#define svmls_n_s64_x(...) __builtin_sve_svmls_n_s64_x(__VA_ARGS__) -#define svmls_n_s16_x(...) __builtin_sve_svmls_n_s16_x(__VA_ARGS__) -#define svmls_n_u8_z(...) __builtin_sve_svmls_n_u8_z(__VA_ARGS__) -#define svmls_n_u32_z(...) __builtin_sve_svmls_n_u32_z(__VA_ARGS__) -#define svmls_n_u64_z(...) __builtin_sve_svmls_n_u64_z(__VA_ARGS__) -#define svmls_n_u16_z(...) __builtin_sve_svmls_n_u16_z(__VA_ARGS__) -#define svmls_n_s8_z(...) __builtin_sve_svmls_n_s8_z(__VA_ARGS__) -#define svmls_n_s32_z(...) __builtin_sve_svmls_n_s32_z(__VA_ARGS__) -#define svmls_n_s64_z(...) __builtin_sve_svmls_n_s64_z(__VA_ARGS__) -#define svmls_n_s16_z(...) __builtin_sve_svmls_n_s16_z(__VA_ARGS__) -#define svmls_f64_m(...) __builtin_sve_svmls_f64_m(__VA_ARGS__) -#define svmls_f32_m(...) __builtin_sve_svmls_f32_m(__VA_ARGS__) -#define svmls_f16_m(...) __builtin_sve_svmls_f16_m(__VA_ARGS__) -#define svmls_f64_x(...) __builtin_sve_svmls_f64_x(__VA_ARGS__) -#define svmls_f32_x(...) __builtin_sve_svmls_f32_x(__VA_ARGS__) -#define svmls_f16_x(...) __builtin_sve_svmls_f16_x(__VA_ARGS__) -#define svmls_f64_z(...) __builtin_sve_svmls_f64_z(__VA_ARGS__) -#define svmls_f32_z(...) __builtin_sve_svmls_f32_z(__VA_ARGS__) -#define svmls_f16_z(...) __builtin_sve_svmls_f16_z(__VA_ARGS__) -#define svmls_u8_m(...) __builtin_sve_svmls_u8_m(__VA_ARGS__) -#define svmls_u32_m(...) __builtin_sve_svmls_u32_m(__VA_ARGS__) -#define svmls_u64_m(...) __builtin_sve_svmls_u64_m(__VA_ARGS__) -#define svmls_u16_m(...) __builtin_sve_svmls_u16_m(__VA_ARGS__) -#define svmls_s8_m(...) __builtin_sve_svmls_s8_m(__VA_ARGS__) -#define svmls_s32_m(...) __builtin_sve_svmls_s32_m(__VA_ARGS__) -#define svmls_s64_m(...) __builtin_sve_svmls_s64_m(__VA_ARGS__) -#define svmls_s16_m(...) __builtin_sve_svmls_s16_m(__VA_ARGS__) -#define svmls_u8_x(...) __builtin_sve_svmls_u8_x(__VA_ARGS__) -#define svmls_u32_x(...) __builtin_sve_svmls_u32_x(__VA_ARGS__) -#define svmls_u64_x(...) __builtin_sve_svmls_u64_x(__VA_ARGS__) -#define svmls_u16_x(...) __builtin_sve_svmls_u16_x(__VA_ARGS__) -#define svmls_s8_x(...) __builtin_sve_svmls_s8_x(__VA_ARGS__) -#define svmls_s32_x(...) __builtin_sve_svmls_s32_x(__VA_ARGS__) -#define svmls_s64_x(...) __builtin_sve_svmls_s64_x(__VA_ARGS__) -#define svmls_s16_x(...) __builtin_sve_svmls_s16_x(__VA_ARGS__) -#define svmls_u8_z(...) __builtin_sve_svmls_u8_z(__VA_ARGS__) -#define svmls_u32_z(...) __builtin_sve_svmls_u32_z(__VA_ARGS__) -#define svmls_u64_z(...) __builtin_sve_svmls_u64_z(__VA_ARGS__) -#define svmls_u16_z(...) __builtin_sve_svmls_u16_z(__VA_ARGS__) -#define svmls_s8_z(...) __builtin_sve_svmls_s8_z(__VA_ARGS__) -#define svmls_s32_z(...) __builtin_sve_svmls_s32_z(__VA_ARGS__) -#define svmls_s64_z(...) __builtin_sve_svmls_s64_z(__VA_ARGS__) -#define svmls_s16_z(...) __builtin_sve_svmls_s16_z(__VA_ARGS__) -#define svmls_lane_f64(...) __builtin_sve_svmls_lane_f64(__VA_ARGS__) -#define svmls_lane_f32(...) __builtin_sve_svmls_lane_f32(__VA_ARGS__) -#define svmls_lane_f16(...) __builtin_sve_svmls_lane_f16(__VA_ARGS__) -#define svmov_b_z(...) __builtin_sve_svmov_b_z(__VA_ARGS__) -#define svmsb_n_f64_m(...) __builtin_sve_svmsb_n_f64_m(__VA_ARGS__) -#define svmsb_n_f32_m(...) __builtin_sve_svmsb_n_f32_m(__VA_ARGS__) -#define svmsb_n_f16_m(...) __builtin_sve_svmsb_n_f16_m(__VA_ARGS__) -#define svmsb_n_f64_x(...) __builtin_sve_svmsb_n_f64_x(__VA_ARGS__) -#define svmsb_n_f32_x(...) __builtin_sve_svmsb_n_f32_x(__VA_ARGS__) -#define svmsb_n_f16_x(...) __builtin_sve_svmsb_n_f16_x(__VA_ARGS__) -#define svmsb_n_f64_z(...) __builtin_sve_svmsb_n_f64_z(__VA_ARGS__) -#define svmsb_n_f32_z(...) __builtin_sve_svmsb_n_f32_z(__VA_ARGS__) -#define svmsb_n_f16_z(...) __builtin_sve_svmsb_n_f16_z(__VA_ARGS__) -#define svmsb_n_u8_m(...) __builtin_sve_svmsb_n_u8_m(__VA_ARGS__) -#define svmsb_n_u32_m(...) __builtin_sve_svmsb_n_u32_m(__VA_ARGS__) -#define svmsb_n_u64_m(...) __builtin_sve_svmsb_n_u64_m(__VA_ARGS__) -#define svmsb_n_u16_m(...) __builtin_sve_svmsb_n_u16_m(__VA_ARGS__) -#define svmsb_n_s8_m(...) __builtin_sve_svmsb_n_s8_m(__VA_ARGS__) -#define svmsb_n_s32_m(...) __builtin_sve_svmsb_n_s32_m(__VA_ARGS__) -#define svmsb_n_s64_m(...) __builtin_sve_svmsb_n_s64_m(__VA_ARGS__) -#define svmsb_n_s16_m(...) __builtin_sve_svmsb_n_s16_m(__VA_ARGS__) -#define svmsb_n_u8_x(...) __builtin_sve_svmsb_n_u8_x(__VA_ARGS__) -#define svmsb_n_u32_x(...) __builtin_sve_svmsb_n_u32_x(__VA_ARGS__) -#define svmsb_n_u64_x(...) __builtin_sve_svmsb_n_u64_x(__VA_ARGS__) -#define svmsb_n_u16_x(...) __builtin_sve_svmsb_n_u16_x(__VA_ARGS__) -#define svmsb_n_s8_x(...) __builtin_sve_svmsb_n_s8_x(__VA_ARGS__) -#define svmsb_n_s32_x(...) __builtin_sve_svmsb_n_s32_x(__VA_ARGS__) -#define svmsb_n_s64_x(...) __builtin_sve_svmsb_n_s64_x(__VA_ARGS__) -#define svmsb_n_s16_x(...) __builtin_sve_svmsb_n_s16_x(__VA_ARGS__) -#define svmsb_n_u8_z(...) __builtin_sve_svmsb_n_u8_z(__VA_ARGS__) -#define svmsb_n_u32_z(...) __builtin_sve_svmsb_n_u32_z(__VA_ARGS__) -#define svmsb_n_u64_z(...) __builtin_sve_svmsb_n_u64_z(__VA_ARGS__) -#define svmsb_n_u16_z(...) __builtin_sve_svmsb_n_u16_z(__VA_ARGS__) -#define svmsb_n_s8_z(...) __builtin_sve_svmsb_n_s8_z(__VA_ARGS__) -#define svmsb_n_s32_z(...) __builtin_sve_svmsb_n_s32_z(__VA_ARGS__) -#define svmsb_n_s64_z(...) __builtin_sve_svmsb_n_s64_z(__VA_ARGS__) -#define svmsb_n_s16_z(...) __builtin_sve_svmsb_n_s16_z(__VA_ARGS__) -#define svmsb_f64_m(...) __builtin_sve_svmsb_f64_m(__VA_ARGS__) -#define svmsb_f32_m(...) __builtin_sve_svmsb_f32_m(__VA_ARGS__) -#define svmsb_f16_m(...) __builtin_sve_svmsb_f16_m(__VA_ARGS__) -#define svmsb_f64_x(...) __builtin_sve_svmsb_f64_x(__VA_ARGS__) -#define svmsb_f32_x(...) __builtin_sve_svmsb_f32_x(__VA_ARGS__) -#define svmsb_f16_x(...) __builtin_sve_svmsb_f16_x(__VA_ARGS__) -#define svmsb_f64_z(...) __builtin_sve_svmsb_f64_z(__VA_ARGS__) -#define svmsb_f32_z(...) __builtin_sve_svmsb_f32_z(__VA_ARGS__) -#define svmsb_f16_z(...) __builtin_sve_svmsb_f16_z(__VA_ARGS__) -#define svmsb_u8_m(...) __builtin_sve_svmsb_u8_m(__VA_ARGS__) -#define svmsb_u32_m(...) __builtin_sve_svmsb_u32_m(__VA_ARGS__) -#define svmsb_u64_m(...) __builtin_sve_svmsb_u64_m(__VA_ARGS__) -#define svmsb_u16_m(...) __builtin_sve_svmsb_u16_m(__VA_ARGS__) -#define svmsb_s8_m(...) __builtin_sve_svmsb_s8_m(__VA_ARGS__) -#define svmsb_s32_m(...) __builtin_sve_svmsb_s32_m(__VA_ARGS__) -#define svmsb_s64_m(...) __builtin_sve_svmsb_s64_m(__VA_ARGS__) -#define svmsb_s16_m(...) __builtin_sve_svmsb_s16_m(__VA_ARGS__) -#define svmsb_u8_x(...) __builtin_sve_svmsb_u8_x(__VA_ARGS__) -#define svmsb_u32_x(...) __builtin_sve_svmsb_u32_x(__VA_ARGS__) -#define svmsb_u64_x(...) __builtin_sve_svmsb_u64_x(__VA_ARGS__) -#define svmsb_u16_x(...) __builtin_sve_svmsb_u16_x(__VA_ARGS__) -#define svmsb_s8_x(...) __builtin_sve_svmsb_s8_x(__VA_ARGS__) -#define svmsb_s32_x(...) __builtin_sve_svmsb_s32_x(__VA_ARGS__) -#define svmsb_s64_x(...) __builtin_sve_svmsb_s64_x(__VA_ARGS__) -#define svmsb_s16_x(...) __builtin_sve_svmsb_s16_x(__VA_ARGS__) -#define svmsb_u8_z(...) __builtin_sve_svmsb_u8_z(__VA_ARGS__) -#define svmsb_u32_z(...) __builtin_sve_svmsb_u32_z(__VA_ARGS__) -#define svmsb_u64_z(...) __builtin_sve_svmsb_u64_z(__VA_ARGS__) -#define svmsb_u16_z(...) __builtin_sve_svmsb_u16_z(__VA_ARGS__) -#define svmsb_s8_z(...) __builtin_sve_svmsb_s8_z(__VA_ARGS__) -#define svmsb_s32_z(...) __builtin_sve_svmsb_s32_z(__VA_ARGS__) -#define svmsb_s64_z(...) __builtin_sve_svmsb_s64_z(__VA_ARGS__) -#define svmsb_s16_z(...) __builtin_sve_svmsb_s16_z(__VA_ARGS__) -#define svmul_n_f64_m(...) __builtin_sve_svmul_n_f64_m(__VA_ARGS__) -#define svmul_n_f32_m(...) __builtin_sve_svmul_n_f32_m(__VA_ARGS__) -#define svmul_n_f16_m(...) __builtin_sve_svmul_n_f16_m(__VA_ARGS__) -#define svmul_n_f64_x(...) __builtin_sve_svmul_n_f64_x(__VA_ARGS__) -#define svmul_n_f32_x(...) __builtin_sve_svmul_n_f32_x(__VA_ARGS__) -#define svmul_n_f16_x(...) __builtin_sve_svmul_n_f16_x(__VA_ARGS__) -#define svmul_n_f64_z(...) __builtin_sve_svmul_n_f64_z(__VA_ARGS__) -#define svmul_n_f32_z(...) __builtin_sve_svmul_n_f32_z(__VA_ARGS__) -#define svmul_n_f16_z(...) __builtin_sve_svmul_n_f16_z(__VA_ARGS__) -#define svmul_n_u8_m(...) __builtin_sve_svmul_n_u8_m(__VA_ARGS__) -#define svmul_n_u32_m(...) __builtin_sve_svmul_n_u32_m(__VA_ARGS__) -#define svmul_n_u64_m(...) __builtin_sve_svmul_n_u64_m(__VA_ARGS__) -#define svmul_n_u16_m(...) __builtin_sve_svmul_n_u16_m(__VA_ARGS__) -#define svmul_n_s8_m(...) __builtin_sve_svmul_n_s8_m(__VA_ARGS__) -#define svmul_n_s32_m(...) __builtin_sve_svmul_n_s32_m(__VA_ARGS__) -#define svmul_n_s64_m(...) __builtin_sve_svmul_n_s64_m(__VA_ARGS__) -#define svmul_n_s16_m(...) __builtin_sve_svmul_n_s16_m(__VA_ARGS__) -#define svmul_n_u8_x(...) __builtin_sve_svmul_n_u8_x(__VA_ARGS__) -#define svmul_n_u32_x(...) __builtin_sve_svmul_n_u32_x(__VA_ARGS__) -#define svmul_n_u64_x(...) __builtin_sve_svmul_n_u64_x(__VA_ARGS__) -#define svmul_n_u16_x(...) __builtin_sve_svmul_n_u16_x(__VA_ARGS__) -#define svmul_n_s8_x(...) __builtin_sve_svmul_n_s8_x(__VA_ARGS__) -#define svmul_n_s32_x(...) __builtin_sve_svmul_n_s32_x(__VA_ARGS__) -#define svmul_n_s64_x(...) __builtin_sve_svmul_n_s64_x(__VA_ARGS__) -#define svmul_n_s16_x(...) __builtin_sve_svmul_n_s16_x(__VA_ARGS__) -#define svmul_n_u8_z(...) __builtin_sve_svmul_n_u8_z(__VA_ARGS__) -#define svmul_n_u32_z(...) __builtin_sve_svmul_n_u32_z(__VA_ARGS__) -#define svmul_n_u64_z(...) __builtin_sve_svmul_n_u64_z(__VA_ARGS__) -#define svmul_n_u16_z(...) __builtin_sve_svmul_n_u16_z(__VA_ARGS__) -#define svmul_n_s8_z(...) __builtin_sve_svmul_n_s8_z(__VA_ARGS__) -#define svmul_n_s32_z(...) __builtin_sve_svmul_n_s32_z(__VA_ARGS__) -#define svmul_n_s64_z(...) __builtin_sve_svmul_n_s64_z(__VA_ARGS__) -#define svmul_n_s16_z(...) __builtin_sve_svmul_n_s16_z(__VA_ARGS__) -#define svmul_f64_m(...) __builtin_sve_svmul_f64_m(__VA_ARGS__) -#define svmul_f32_m(...) __builtin_sve_svmul_f32_m(__VA_ARGS__) -#define svmul_f16_m(...) __builtin_sve_svmul_f16_m(__VA_ARGS__) -#define svmul_f64_x(...) __builtin_sve_svmul_f64_x(__VA_ARGS__) -#define svmul_f32_x(...) __builtin_sve_svmul_f32_x(__VA_ARGS__) -#define svmul_f16_x(...) __builtin_sve_svmul_f16_x(__VA_ARGS__) -#define svmul_f64_z(...) __builtin_sve_svmul_f64_z(__VA_ARGS__) -#define svmul_f32_z(...) __builtin_sve_svmul_f32_z(__VA_ARGS__) -#define svmul_f16_z(...) __builtin_sve_svmul_f16_z(__VA_ARGS__) -#define svmul_u8_m(...) __builtin_sve_svmul_u8_m(__VA_ARGS__) -#define svmul_u32_m(...) __builtin_sve_svmul_u32_m(__VA_ARGS__) -#define svmul_u64_m(...) __builtin_sve_svmul_u64_m(__VA_ARGS__) -#define svmul_u16_m(...) __builtin_sve_svmul_u16_m(__VA_ARGS__) -#define svmul_s8_m(...) __builtin_sve_svmul_s8_m(__VA_ARGS__) -#define svmul_s32_m(...) __builtin_sve_svmul_s32_m(__VA_ARGS__) -#define svmul_s64_m(...) __builtin_sve_svmul_s64_m(__VA_ARGS__) -#define svmul_s16_m(...) __builtin_sve_svmul_s16_m(__VA_ARGS__) -#define svmul_u8_x(...) __builtin_sve_svmul_u8_x(__VA_ARGS__) -#define svmul_u32_x(...) __builtin_sve_svmul_u32_x(__VA_ARGS__) -#define svmul_u64_x(...) __builtin_sve_svmul_u64_x(__VA_ARGS__) -#define svmul_u16_x(...) __builtin_sve_svmul_u16_x(__VA_ARGS__) -#define svmul_s8_x(...) __builtin_sve_svmul_s8_x(__VA_ARGS__) -#define svmul_s32_x(...) __builtin_sve_svmul_s32_x(__VA_ARGS__) -#define svmul_s64_x(...) __builtin_sve_svmul_s64_x(__VA_ARGS__) -#define svmul_s16_x(...) __builtin_sve_svmul_s16_x(__VA_ARGS__) -#define svmul_u8_z(...) __builtin_sve_svmul_u8_z(__VA_ARGS__) -#define svmul_u32_z(...) __builtin_sve_svmul_u32_z(__VA_ARGS__) -#define svmul_u64_z(...) __builtin_sve_svmul_u64_z(__VA_ARGS__) -#define svmul_u16_z(...) __builtin_sve_svmul_u16_z(__VA_ARGS__) -#define svmul_s8_z(...) __builtin_sve_svmul_s8_z(__VA_ARGS__) -#define svmul_s32_z(...) __builtin_sve_svmul_s32_z(__VA_ARGS__) -#define svmul_s64_z(...) __builtin_sve_svmul_s64_z(__VA_ARGS__) -#define svmul_s16_z(...) __builtin_sve_svmul_s16_z(__VA_ARGS__) -#define svmul_lane_f64(...) __builtin_sve_svmul_lane_f64(__VA_ARGS__) -#define svmul_lane_f32(...) __builtin_sve_svmul_lane_f32(__VA_ARGS__) -#define svmul_lane_f16(...) __builtin_sve_svmul_lane_f16(__VA_ARGS__) -#define svmulh_n_s8_m(...) __builtin_sve_svmulh_n_s8_m(__VA_ARGS__) -#define svmulh_n_s32_m(...) __builtin_sve_svmulh_n_s32_m(__VA_ARGS__) -#define svmulh_n_s64_m(...) __builtin_sve_svmulh_n_s64_m(__VA_ARGS__) -#define svmulh_n_s16_m(...) __builtin_sve_svmulh_n_s16_m(__VA_ARGS__) -#define svmulh_n_s8_x(...) __builtin_sve_svmulh_n_s8_x(__VA_ARGS__) -#define svmulh_n_s32_x(...) __builtin_sve_svmulh_n_s32_x(__VA_ARGS__) -#define svmulh_n_s64_x(...) __builtin_sve_svmulh_n_s64_x(__VA_ARGS__) -#define svmulh_n_s16_x(...) __builtin_sve_svmulh_n_s16_x(__VA_ARGS__) -#define svmulh_n_s8_z(...) __builtin_sve_svmulh_n_s8_z(__VA_ARGS__) -#define svmulh_n_s32_z(...) __builtin_sve_svmulh_n_s32_z(__VA_ARGS__) -#define svmulh_n_s64_z(...) __builtin_sve_svmulh_n_s64_z(__VA_ARGS__) -#define svmulh_n_s16_z(...) __builtin_sve_svmulh_n_s16_z(__VA_ARGS__) -#define svmulh_n_u8_m(...) __builtin_sve_svmulh_n_u8_m(__VA_ARGS__) -#define svmulh_n_u32_m(...) __builtin_sve_svmulh_n_u32_m(__VA_ARGS__) -#define svmulh_n_u64_m(...) __builtin_sve_svmulh_n_u64_m(__VA_ARGS__) -#define svmulh_n_u16_m(...) __builtin_sve_svmulh_n_u16_m(__VA_ARGS__) -#define svmulh_n_u8_x(...) __builtin_sve_svmulh_n_u8_x(__VA_ARGS__) -#define svmulh_n_u32_x(...) __builtin_sve_svmulh_n_u32_x(__VA_ARGS__) -#define svmulh_n_u64_x(...) __builtin_sve_svmulh_n_u64_x(__VA_ARGS__) -#define svmulh_n_u16_x(...) __builtin_sve_svmulh_n_u16_x(__VA_ARGS__) -#define svmulh_n_u8_z(...) __builtin_sve_svmulh_n_u8_z(__VA_ARGS__) -#define svmulh_n_u32_z(...) __builtin_sve_svmulh_n_u32_z(__VA_ARGS__) -#define svmulh_n_u64_z(...) __builtin_sve_svmulh_n_u64_z(__VA_ARGS__) -#define svmulh_n_u16_z(...) __builtin_sve_svmulh_n_u16_z(__VA_ARGS__) -#define svmulh_s8_m(...) __builtin_sve_svmulh_s8_m(__VA_ARGS__) -#define svmulh_s32_m(...) __builtin_sve_svmulh_s32_m(__VA_ARGS__) -#define svmulh_s64_m(...) __builtin_sve_svmulh_s64_m(__VA_ARGS__) -#define svmulh_s16_m(...) __builtin_sve_svmulh_s16_m(__VA_ARGS__) -#define svmulh_s8_x(...) __builtin_sve_svmulh_s8_x(__VA_ARGS__) -#define svmulh_s32_x(...) __builtin_sve_svmulh_s32_x(__VA_ARGS__) -#define svmulh_s64_x(...) __builtin_sve_svmulh_s64_x(__VA_ARGS__) -#define svmulh_s16_x(...) __builtin_sve_svmulh_s16_x(__VA_ARGS__) -#define svmulh_s8_z(...) __builtin_sve_svmulh_s8_z(__VA_ARGS__) -#define svmulh_s32_z(...) __builtin_sve_svmulh_s32_z(__VA_ARGS__) -#define svmulh_s64_z(...) __builtin_sve_svmulh_s64_z(__VA_ARGS__) -#define svmulh_s16_z(...) __builtin_sve_svmulh_s16_z(__VA_ARGS__) -#define svmulh_u8_m(...) __builtin_sve_svmulh_u8_m(__VA_ARGS__) -#define svmulh_u32_m(...) __builtin_sve_svmulh_u32_m(__VA_ARGS__) -#define svmulh_u64_m(...) __builtin_sve_svmulh_u64_m(__VA_ARGS__) -#define svmulh_u16_m(...) __builtin_sve_svmulh_u16_m(__VA_ARGS__) -#define svmulh_u8_x(...) __builtin_sve_svmulh_u8_x(__VA_ARGS__) -#define svmulh_u32_x(...) __builtin_sve_svmulh_u32_x(__VA_ARGS__) -#define svmulh_u64_x(...) __builtin_sve_svmulh_u64_x(__VA_ARGS__) -#define svmulh_u16_x(...) __builtin_sve_svmulh_u16_x(__VA_ARGS__) -#define svmulh_u8_z(...) __builtin_sve_svmulh_u8_z(__VA_ARGS__) -#define svmulh_u32_z(...) __builtin_sve_svmulh_u32_z(__VA_ARGS__) -#define svmulh_u64_z(...) __builtin_sve_svmulh_u64_z(__VA_ARGS__) -#define svmulh_u16_z(...) __builtin_sve_svmulh_u16_z(__VA_ARGS__) -#define svmulx_n_f64_m(...) __builtin_sve_svmulx_n_f64_m(__VA_ARGS__) -#define svmulx_n_f32_m(...) __builtin_sve_svmulx_n_f32_m(__VA_ARGS__) -#define svmulx_n_f16_m(...) __builtin_sve_svmulx_n_f16_m(__VA_ARGS__) -#define svmulx_n_f64_x(...) __builtin_sve_svmulx_n_f64_x(__VA_ARGS__) -#define svmulx_n_f32_x(...) __builtin_sve_svmulx_n_f32_x(__VA_ARGS__) -#define svmulx_n_f16_x(...) __builtin_sve_svmulx_n_f16_x(__VA_ARGS__) -#define svmulx_n_f64_z(...) __builtin_sve_svmulx_n_f64_z(__VA_ARGS__) -#define svmulx_n_f32_z(...) __builtin_sve_svmulx_n_f32_z(__VA_ARGS__) -#define svmulx_n_f16_z(...) __builtin_sve_svmulx_n_f16_z(__VA_ARGS__) -#define svmulx_f64_m(...) __builtin_sve_svmulx_f64_m(__VA_ARGS__) -#define svmulx_f32_m(...) __builtin_sve_svmulx_f32_m(__VA_ARGS__) -#define svmulx_f16_m(...) __builtin_sve_svmulx_f16_m(__VA_ARGS__) -#define svmulx_f64_x(...) __builtin_sve_svmulx_f64_x(__VA_ARGS__) -#define svmulx_f32_x(...) __builtin_sve_svmulx_f32_x(__VA_ARGS__) -#define svmulx_f16_x(...) __builtin_sve_svmulx_f16_x(__VA_ARGS__) -#define svmulx_f64_z(...) __builtin_sve_svmulx_f64_z(__VA_ARGS__) -#define svmulx_f32_z(...) __builtin_sve_svmulx_f32_z(__VA_ARGS__) -#define svmulx_f16_z(...) __builtin_sve_svmulx_f16_z(__VA_ARGS__) -#define svnand_b_z(...) __builtin_sve_svnand_b_z(__VA_ARGS__) -#define svneg_f64_m(...) __builtin_sve_svneg_f64_m(__VA_ARGS__) -#define svneg_f32_m(...) __builtin_sve_svneg_f32_m(__VA_ARGS__) -#define svneg_f16_m(...) __builtin_sve_svneg_f16_m(__VA_ARGS__) -#define svneg_f64_x(...) __builtin_sve_svneg_f64_x(__VA_ARGS__) -#define svneg_f32_x(...) __builtin_sve_svneg_f32_x(__VA_ARGS__) -#define svneg_f16_x(...) __builtin_sve_svneg_f16_x(__VA_ARGS__) -#define svneg_f64_z(...) __builtin_sve_svneg_f64_z(__VA_ARGS__) -#define svneg_f32_z(...) __builtin_sve_svneg_f32_z(__VA_ARGS__) -#define svneg_f16_z(...) __builtin_sve_svneg_f16_z(__VA_ARGS__) -#define svneg_s8_m(...) __builtin_sve_svneg_s8_m(__VA_ARGS__) -#define svneg_s32_m(...) __builtin_sve_svneg_s32_m(__VA_ARGS__) -#define svneg_s64_m(...) __builtin_sve_svneg_s64_m(__VA_ARGS__) -#define svneg_s16_m(...) __builtin_sve_svneg_s16_m(__VA_ARGS__) -#define svneg_s8_x(...) __builtin_sve_svneg_s8_x(__VA_ARGS__) -#define svneg_s32_x(...) __builtin_sve_svneg_s32_x(__VA_ARGS__) -#define svneg_s64_x(...) __builtin_sve_svneg_s64_x(__VA_ARGS__) -#define svneg_s16_x(...) __builtin_sve_svneg_s16_x(__VA_ARGS__) -#define svneg_s8_z(...) __builtin_sve_svneg_s8_z(__VA_ARGS__) -#define svneg_s32_z(...) __builtin_sve_svneg_s32_z(__VA_ARGS__) -#define svneg_s64_z(...) __builtin_sve_svneg_s64_z(__VA_ARGS__) -#define svneg_s16_z(...) __builtin_sve_svneg_s16_z(__VA_ARGS__) -#define svnmad_n_f64_m(...) __builtin_sve_svnmad_n_f64_m(__VA_ARGS__) -#define svnmad_n_f32_m(...) __builtin_sve_svnmad_n_f32_m(__VA_ARGS__) -#define svnmad_n_f16_m(...) __builtin_sve_svnmad_n_f16_m(__VA_ARGS__) -#define svnmad_n_f64_x(...) __builtin_sve_svnmad_n_f64_x(__VA_ARGS__) -#define svnmad_n_f32_x(...) __builtin_sve_svnmad_n_f32_x(__VA_ARGS__) -#define svnmad_n_f16_x(...) __builtin_sve_svnmad_n_f16_x(__VA_ARGS__) -#define svnmad_n_f64_z(...) __builtin_sve_svnmad_n_f64_z(__VA_ARGS__) -#define svnmad_n_f32_z(...) __builtin_sve_svnmad_n_f32_z(__VA_ARGS__) -#define svnmad_n_f16_z(...) __builtin_sve_svnmad_n_f16_z(__VA_ARGS__) -#define svnmad_f64_m(...) __builtin_sve_svnmad_f64_m(__VA_ARGS__) -#define svnmad_f32_m(...) __builtin_sve_svnmad_f32_m(__VA_ARGS__) -#define svnmad_f16_m(...) __builtin_sve_svnmad_f16_m(__VA_ARGS__) -#define svnmad_f64_x(...) __builtin_sve_svnmad_f64_x(__VA_ARGS__) -#define svnmad_f32_x(...) __builtin_sve_svnmad_f32_x(__VA_ARGS__) -#define svnmad_f16_x(...) __builtin_sve_svnmad_f16_x(__VA_ARGS__) -#define svnmad_f64_z(...) __builtin_sve_svnmad_f64_z(__VA_ARGS__) -#define svnmad_f32_z(...) __builtin_sve_svnmad_f32_z(__VA_ARGS__) -#define svnmad_f16_z(...) __builtin_sve_svnmad_f16_z(__VA_ARGS__) -#define svnmla_n_f64_m(...) __builtin_sve_svnmla_n_f64_m(__VA_ARGS__) -#define svnmla_n_f32_m(...) __builtin_sve_svnmla_n_f32_m(__VA_ARGS__) -#define svnmla_n_f16_m(...) __builtin_sve_svnmla_n_f16_m(__VA_ARGS__) -#define svnmla_n_f64_x(...) __builtin_sve_svnmla_n_f64_x(__VA_ARGS__) -#define svnmla_n_f32_x(...) __builtin_sve_svnmla_n_f32_x(__VA_ARGS__) -#define svnmla_n_f16_x(...) __builtin_sve_svnmla_n_f16_x(__VA_ARGS__) -#define svnmla_n_f64_z(...) __builtin_sve_svnmla_n_f64_z(__VA_ARGS__) -#define svnmla_n_f32_z(...) __builtin_sve_svnmla_n_f32_z(__VA_ARGS__) -#define svnmla_n_f16_z(...) __builtin_sve_svnmla_n_f16_z(__VA_ARGS__) -#define svnmla_f64_m(...) __builtin_sve_svnmla_f64_m(__VA_ARGS__) -#define svnmla_f32_m(...) __builtin_sve_svnmla_f32_m(__VA_ARGS__) -#define svnmla_f16_m(...) __builtin_sve_svnmla_f16_m(__VA_ARGS__) -#define svnmla_f64_x(...) __builtin_sve_svnmla_f64_x(__VA_ARGS__) -#define svnmla_f32_x(...) __builtin_sve_svnmla_f32_x(__VA_ARGS__) -#define svnmla_f16_x(...) __builtin_sve_svnmla_f16_x(__VA_ARGS__) -#define svnmla_f64_z(...) __builtin_sve_svnmla_f64_z(__VA_ARGS__) -#define svnmla_f32_z(...) __builtin_sve_svnmla_f32_z(__VA_ARGS__) -#define svnmla_f16_z(...) __builtin_sve_svnmla_f16_z(__VA_ARGS__) -#define svnmls_n_f64_m(...) __builtin_sve_svnmls_n_f64_m(__VA_ARGS__) -#define svnmls_n_f32_m(...) __builtin_sve_svnmls_n_f32_m(__VA_ARGS__) -#define svnmls_n_f16_m(...) __builtin_sve_svnmls_n_f16_m(__VA_ARGS__) -#define svnmls_n_f64_x(...) __builtin_sve_svnmls_n_f64_x(__VA_ARGS__) -#define svnmls_n_f32_x(...) __builtin_sve_svnmls_n_f32_x(__VA_ARGS__) -#define svnmls_n_f16_x(...) __builtin_sve_svnmls_n_f16_x(__VA_ARGS__) -#define svnmls_n_f64_z(...) __builtin_sve_svnmls_n_f64_z(__VA_ARGS__) -#define svnmls_n_f32_z(...) __builtin_sve_svnmls_n_f32_z(__VA_ARGS__) -#define svnmls_n_f16_z(...) __builtin_sve_svnmls_n_f16_z(__VA_ARGS__) -#define svnmls_f64_m(...) __builtin_sve_svnmls_f64_m(__VA_ARGS__) -#define svnmls_f32_m(...) __builtin_sve_svnmls_f32_m(__VA_ARGS__) -#define svnmls_f16_m(...) __builtin_sve_svnmls_f16_m(__VA_ARGS__) -#define svnmls_f64_x(...) __builtin_sve_svnmls_f64_x(__VA_ARGS__) -#define svnmls_f32_x(...) __builtin_sve_svnmls_f32_x(__VA_ARGS__) -#define svnmls_f16_x(...) __builtin_sve_svnmls_f16_x(__VA_ARGS__) -#define svnmls_f64_z(...) __builtin_sve_svnmls_f64_z(__VA_ARGS__) -#define svnmls_f32_z(...) __builtin_sve_svnmls_f32_z(__VA_ARGS__) -#define svnmls_f16_z(...) __builtin_sve_svnmls_f16_z(__VA_ARGS__) -#define svnmsb_n_f64_m(...) __builtin_sve_svnmsb_n_f64_m(__VA_ARGS__) -#define svnmsb_n_f32_m(...) __builtin_sve_svnmsb_n_f32_m(__VA_ARGS__) -#define svnmsb_n_f16_m(...) __builtin_sve_svnmsb_n_f16_m(__VA_ARGS__) -#define svnmsb_n_f64_x(...) __builtin_sve_svnmsb_n_f64_x(__VA_ARGS__) -#define svnmsb_n_f32_x(...) __builtin_sve_svnmsb_n_f32_x(__VA_ARGS__) -#define svnmsb_n_f16_x(...) __builtin_sve_svnmsb_n_f16_x(__VA_ARGS__) -#define svnmsb_n_f64_z(...) __builtin_sve_svnmsb_n_f64_z(__VA_ARGS__) -#define svnmsb_n_f32_z(...) __builtin_sve_svnmsb_n_f32_z(__VA_ARGS__) -#define svnmsb_n_f16_z(...) __builtin_sve_svnmsb_n_f16_z(__VA_ARGS__) -#define svnmsb_f64_m(...) __builtin_sve_svnmsb_f64_m(__VA_ARGS__) -#define svnmsb_f32_m(...) __builtin_sve_svnmsb_f32_m(__VA_ARGS__) -#define svnmsb_f16_m(...) __builtin_sve_svnmsb_f16_m(__VA_ARGS__) -#define svnmsb_f64_x(...) __builtin_sve_svnmsb_f64_x(__VA_ARGS__) -#define svnmsb_f32_x(...) __builtin_sve_svnmsb_f32_x(__VA_ARGS__) -#define svnmsb_f16_x(...) __builtin_sve_svnmsb_f16_x(__VA_ARGS__) -#define svnmsb_f64_z(...) __builtin_sve_svnmsb_f64_z(__VA_ARGS__) -#define svnmsb_f32_z(...) __builtin_sve_svnmsb_f32_z(__VA_ARGS__) -#define svnmsb_f16_z(...) __builtin_sve_svnmsb_f16_z(__VA_ARGS__) -#define svnor_b_z(...) __builtin_sve_svnor_b_z(__VA_ARGS__) -#define svnot_b_z(...) __builtin_sve_svnot_b_z(__VA_ARGS__) -#define svnot_u8_m(...) __builtin_sve_svnot_u8_m(__VA_ARGS__) -#define svnot_u32_m(...) __builtin_sve_svnot_u32_m(__VA_ARGS__) -#define svnot_u64_m(...) __builtin_sve_svnot_u64_m(__VA_ARGS__) -#define svnot_u16_m(...) __builtin_sve_svnot_u16_m(__VA_ARGS__) -#define svnot_s8_m(...) __builtin_sve_svnot_s8_m(__VA_ARGS__) -#define svnot_s32_m(...) __builtin_sve_svnot_s32_m(__VA_ARGS__) -#define svnot_s64_m(...) __builtin_sve_svnot_s64_m(__VA_ARGS__) -#define svnot_s16_m(...) __builtin_sve_svnot_s16_m(__VA_ARGS__) -#define svnot_u8_x(...) __builtin_sve_svnot_u8_x(__VA_ARGS__) -#define svnot_u32_x(...) __builtin_sve_svnot_u32_x(__VA_ARGS__) -#define svnot_u64_x(...) __builtin_sve_svnot_u64_x(__VA_ARGS__) -#define svnot_u16_x(...) __builtin_sve_svnot_u16_x(__VA_ARGS__) -#define svnot_s8_x(...) __builtin_sve_svnot_s8_x(__VA_ARGS__) -#define svnot_s32_x(...) __builtin_sve_svnot_s32_x(__VA_ARGS__) -#define svnot_s64_x(...) __builtin_sve_svnot_s64_x(__VA_ARGS__) -#define svnot_s16_x(...) __builtin_sve_svnot_s16_x(__VA_ARGS__) -#define svnot_u8_z(...) __builtin_sve_svnot_u8_z(__VA_ARGS__) -#define svnot_u32_z(...) __builtin_sve_svnot_u32_z(__VA_ARGS__) -#define svnot_u64_z(...) __builtin_sve_svnot_u64_z(__VA_ARGS__) -#define svnot_u16_z(...) __builtin_sve_svnot_u16_z(__VA_ARGS__) -#define svnot_s8_z(...) __builtin_sve_svnot_s8_z(__VA_ARGS__) -#define svnot_s32_z(...) __builtin_sve_svnot_s32_z(__VA_ARGS__) -#define svnot_s64_z(...) __builtin_sve_svnot_s64_z(__VA_ARGS__) -#define svnot_s16_z(...) __builtin_sve_svnot_s16_z(__VA_ARGS__) -#define svorn_b_z(...) __builtin_sve_svorn_b_z(__VA_ARGS__) -#define svorr_b_z(...) __builtin_sve_svorr_b_z(__VA_ARGS__) -#define svorr_n_u8_m(...) __builtin_sve_svorr_n_u8_m(__VA_ARGS__) -#define svorr_n_u32_m(...) __builtin_sve_svorr_n_u32_m(__VA_ARGS__) -#define svorr_n_u64_m(...) __builtin_sve_svorr_n_u64_m(__VA_ARGS__) -#define svorr_n_u16_m(...) __builtin_sve_svorr_n_u16_m(__VA_ARGS__) -#define svorr_n_s8_m(...) __builtin_sve_svorr_n_s8_m(__VA_ARGS__) -#define svorr_n_s32_m(...) __builtin_sve_svorr_n_s32_m(__VA_ARGS__) -#define svorr_n_s64_m(...) __builtin_sve_svorr_n_s64_m(__VA_ARGS__) -#define svorr_n_s16_m(...) __builtin_sve_svorr_n_s16_m(__VA_ARGS__) -#define svorr_n_u8_x(...) __builtin_sve_svorr_n_u8_x(__VA_ARGS__) -#define svorr_n_u32_x(...) __builtin_sve_svorr_n_u32_x(__VA_ARGS__) -#define svorr_n_u64_x(...) __builtin_sve_svorr_n_u64_x(__VA_ARGS__) -#define svorr_n_u16_x(...) __builtin_sve_svorr_n_u16_x(__VA_ARGS__) -#define svorr_n_s8_x(...) __builtin_sve_svorr_n_s8_x(__VA_ARGS__) -#define svorr_n_s32_x(...) __builtin_sve_svorr_n_s32_x(__VA_ARGS__) -#define svorr_n_s64_x(...) __builtin_sve_svorr_n_s64_x(__VA_ARGS__) -#define svorr_n_s16_x(...) __builtin_sve_svorr_n_s16_x(__VA_ARGS__) -#define svorr_n_u8_z(...) __builtin_sve_svorr_n_u8_z(__VA_ARGS__) -#define svorr_n_u32_z(...) __builtin_sve_svorr_n_u32_z(__VA_ARGS__) -#define svorr_n_u64_z(...) __builtin_sve_svorr_n_u64_z(__VA_ARGS__) -#define svorr_n_u16_z(...) __builtin_sve_svorr_n_u16_z(__VA_ARGS__) -#define svorr_n_s8_z(...) __builtin_sve_svorr_n_s8_z(__VA_ARGS__) -#define svorr_n_s32_z(...) __builtin_sve_svorr_n_s32_z(__VA_ARGS__) -#define svorr_n_s64_z(...) __builtin_sve_svorr_n_s64_z(__VA_ARGS__) -#define svorr_n_s16_z(...) __builtin_sve_svorr_n_s16_z(__VA_ARGS__) -#define svorr_u8_m(...) __builtin_sve_svorr_u8_m(__VA_ARGS__) -#define svorr_u32_m(...) __builtin_sve_svorr_u32_m(__VA_ARGS__) -#define svorr_u64_m(...) __builtin_sve_svorr_u64_m(__VA_ARGS__) -#define svorr_u16_m(...) __builtin_sve_svorr_u16_m(__VA_ARGS__) -#define svorr_s8_m(...) __builtin_sve_svorr_s8_m(__VA_ARGS__) -#define svorr_s32_m(...) __builtin_sve_svorr_s32_m(__VA_ARGS__) -#define svorr_s64_m(...) __builtin_sve_svorr_s64_m(__VA_ARGS__) -#define svorr_s16_m(...) __builtin_sve_svorr_s16_m(__VA_ARGS__) -#define svorr_u8_x(...) __builtin_sve_svorr_u8_x(__VA_ARGS__) -#define svorr_u32_x(...) __builtin_sve_svorr_u32_x(__VA_ARGS__) -#define svorr_u64_x(...) __builtin_sve_svorr_u64_x(__VA_ARGS__) -#define svorr_u16_x(...) __builtin_sve_svorr_u16_x(__VA_ARGS__) -#define svorr_s8_x(...) __builtin_sve_svorr_s8_x(__VA_ARGS__) -#define svorr_s32_x(...) __builtin_sve_svorr_s32_x(__VA_ARGS__) -#define svorr_s64_x(...) __builtin_sve_svorr_s64_x(__VA_ARGS__) -#define svorr_s16_x(...) __builtin_sve_svorr_s16_x(__VA_ARGS__) -#define svorr_u8_z(...) __builtin_sve_svorr_u8_z(__VA_ARGS__) -#define svorr_u32_z(...) __builtin_sve_svorr_u32_z(__VA_ARGS__) -#define svorr_u64_z(...) __builtin_sve_svorr_u64_z(__VA_ARGS__) -#define svorr_u16_z(...) __builtin_sve_svorr_u16_z(__VA_ARGS__) -#define svorr_s8_z(...) __builtin_sve_svorr_s8_z(__VA_ARGS__) -#define svorr_s32_z(...) __builtin_sve_svorr_s32_z(__VA_ARGS__) -#define svorr_s64_z(...) __builtin_sve_svorr_s64_z(__VA_ARGS__) -#define svorr_s16_z(...) __builtin_sve_svorr_s16_z(__VA_ARGS__) -#define svorv_u8(...) __builtin_sve_svorv_u8(__VA_ARGS__) -#define svorv_u32(...) __builtin_sve_svorv_u32(__VA_ARGS__) -#define svorv_u64(...) __builtin_sve_svorv_u64(__VA_ARGS__) -#define svorv_u16(...) __builtin_sve_svorv_u16(__VA_ARGS__) -#define svorv_s8(...) __builtin_sve_svorv_s8(__VA_ARGS__) -#define svorv_s32(...) __builtin_sve_svorv_s32(__VA_ARGS__) -#define svorv_s64(...) __builtin_sve_svorv_s64(__VA_ARGS__) -#define svorv_s16(...) __builtin_sve_svorv_s16(__VA_ARGS__) -#define svpfalse_b(...) __builtin_sve_svpfalse_b(__VA_ARGS__) -#define svpfirst_b(...) __builtin_sve_svpfirst_b(__VA_ARGS__) -#define svpnext_b8(...) __builtin_sve_svpnext_b8(__VA_ARGS__) -#define svpnext_b32(...) __builtin_sve_svpnext_b32(__VA_ARGS__) -#define svpnext_b64(...) __builtin_sve_svpnext_b64(__VA_ARGS__) -#define svpnext_b16(...) __builtin_sve_svpnext_b16(__VA_ARGS__) -#define svprfb(...) __builtin_sve_svprfb(__VA_ARGS__) -#define svprfb_gather_u32base(...) __builtin_sve_svprfb_gather_u32base(__VA_ARGS__) -#define svprfb_gather_u64base(...) __builtin_sve_svprfb_gather_u64base(__VA_ARGS__) -#define svprfb_gather_u32base_offset(...) __builtin_sve_svprfb_gather_u32base_offset(__VA_ARGS__) -#define svprfb_gather_u64base_offset(...) __builtin_sve_svprfb_gather_u64base_offset(__VA_ARGS__) -#define svprfb_gather_s32offset(...) __builtin_sve_svprfb_gather_s32offset(__VA_ARGS__) -#define svprfb_gather_u32offset(...) __builtin_sve_svprfb_gather_u32offset(__VA_ARGS__) -#define svprfb_gather_s64offset(...) __builtin_sve_svprfb_gather_s64offset(__VA_ARGS__) -#define svprfb_gather_u64offset(...) __builtin_sve_svprfb_gather_u64offset(__VA_ARGS__) -#define svprfb_vnum(...) __builtin_sve_svprfb_vnum(__VA_ARGS__) -#define svprfd(...) __builtin_sve_svprfd(__VA_ARGS__) -#define svprfd_gather_u32base(...) __builtin_sve_svprfd_gather_u32base(__VA_ARGS__) -#define svprfd_gather_u64base(...) __builtin_sve_svprfd_gather_u64base(__VA_ARGS__) -#define svprfd_gather_u32base_index(...) __builtin_sve_svprfd_gather_u32base_index(__VA_ARGS__) -#define svprfd_gather_u64base_index(...) __builtin_sve_svprfd_gather_u64base_index(__VA_ARGS__) -#define svprfd_gather_s32index(...) __builtin_sve_svprfd_gather_s32index(__VA_ARGS__) -#define svprfd_gather_u32index(...) __builtin_sve_svprfd_gather_u32index(__VA_ARGS__) -#define svprfd_gather_s64index(...) __builtin_sve_svprfd_gather_s64index(__VA_ARGS__) -#define svprfd_gather_u64index(...) __builtin_sve_svprfd_gather_u64index(__VA_ARGS__) -#define svprfd_vnum(...) __builtin_sve_svprfd_vnum(__VA_ARGS__) -#define svprfh(...) __builtin_sve_svprfh(__VA_ARGS__) -#define svprfh_gather_u32base(...) __builtin_sve_svprfh_gather_u32base(__VA_ARGS__) -#define svprfh_gather_u64base(...) __builtin_sve_svprfh_gather_u64base(__VA_ARGS__) -#define svprfh_gather_u32base_index(...) __builtin_sve_svprfh_gather_u32base_index(__VA_ARGS__) -#define svprfh_gather_u64base_index(...) __builtin_sve_svprfh_gather_u64base_index(__VA_ARGS__) -#define svprfh_gather_s32index(...) __builtin_sve_svprfh_gather_s32index(__VA_ARGS__) -#define svprfh_gather_u32index(...) __builtin_sve_svprfh_gather_u32index(__VA_ARGS__) -#define svprfh_gather_s64index(...) __builtin_sve_svprfh_gather_s64index(__VA_ARGS__) -#define svprfh_gather_u64index(...) __builtin_sve_svprfh_gather_u64index(__VA_ARGS__) -#define svprfh_vnum(...) __builtin_sve_svprfh_vnum(__VA_ARGS__) -#define svprfw(...) __builtin_sve_svprfw(__VA_ARGS__) -#define svprfw_gather_u32base(...) __builtin_sve_svprfw_gather_u32base(__VA_ARGS__) -#define svprfw_gather_u64base(...) __builtin_sve_svprfw_gather_u64base(__VA_ARGS__) -#define svprfw_gather_u32base_index(...) __builtin_sve_svprfw_gather_u32base_index(__VA_ARGS__) -#define svprfw_gather_u64base_index(...) __builtin_sve_svprfw_gather_u64base_index(__VA_ARGS__) -#define svprfw_gather_s32index(...) __builtin_sve_svprfw_gather_s32index(__VA_ARGS__) -#define svprfw_gather_u32index(...) __builtin_sve_svprfw_gather_u32index(__VA_ARGS__) -#define svprfw_gather_s64index(...) __builtin_sve_svprfw_gather_s64index(__VA_ARGS__) -#define svprfw_gather_u64index(...) __builtin_sve_svprfw_gather_u64index(__VA_ARGS__) -#define svprfw_vnum(...) __builtin_sve_svprfw_vnum(__VA_ARGS__) -#define svptest_any(...) __builtin_sve_svptest_any(__VA_ARGS__) -#define svptest_first(...) __builtin_sve_svptest_first(__VA_ARGS__) -#define svptest_last(...) __builtin_sve_svptest_last(__VA_ARGS__) -#define svptrue_pat_b8(...) __builtin_sve_svptrue_pat_b8(__VA_ARGS__) -#define svptrue_pat_b32(...) __builtin_sve_svptrue_pat_b32(__VA_ARGS__) -#define svptrue_pat_b64(...) __builtin_sve_svptrue_pat_b64(__VA_ARGS__) -#define svptrue_pat_b16(...) __builtin_sve_svptrue_pat_b16(__VA_ARGS__) -#define svptrue_b8(...) __builtin_sve_svptrue_b8(__VA_ARGS__) -#define svptrue_b32(...) __builtin_sve_svptrue_b32(__VA_ARGS__) -#define svptrue_b64(...) __builtin_sve_svptrue_b64(__VA_ARGS__) -#define svptrue_b16(...) __builtin_sve_svptrue_b16(__VA_ARGS__) -#define svqadd_n_s8(...) __builtin_sve_svqadd_n_s8(__VA_ARGS__) -#define svqadd_n_s32(...) __builtin_sve_svqadd_n_s32(__VA_ARGS__) -#define svqadd_n_s64(...) __builtin_sve_svqadd_n_s64(__VA_ARGS__) -#define svqadd_n_s16(...) __builtin_sve_svqadd_n_s16(__VA_ARGS__) -#define svqadd_n_u8(...) __builtin_sve_svqadd_n_u8(__VA_ARGS__) -#define svqadd_n_u32(...) __builtin_sve_svqadd_n_u32(__VA_ARGS__) -#define svqadd_n_u64(...) __builtin_sve_svqadd_n_u64(__VA_ARGS__) -#define svqadd_n_u16(...) __builtin_sve_svqadd_n_u16(__VA_ARGS__) -#define svqadd_s8(...) __builtin_sve_svqadd_s8(__VA_ARGS__) -#define svqadd_s32(...) __builtin_sve_svqadd_s32(__VA_ARGS__) -#define svqadd_s64(...) __builtin_sve_svqadd_s64(__VA_ARGS__) -#define svqadd_s16(...) __builtin_sve_svqadd_s16(__VA_ARGS__) -#define svqadd_u8(...) __builtin_sve_svqadd_u8(__VA_ARGS__) -#define svqadd_u32(...) __builtin_sve_svqadd_u32(__VA_ARGS__) -#define svqadd_u64(...) __builtin_sve_svqadd_u64(__VA_ARGS__) -#define svqadd_u16(...) __builtin_sve_svqadd_u16(__VA_ARGS__) -#define svqdecb_n_s32(...) __builtin_sve_svqdecb_n_s32(__VA_ARGS__) -#define svqdecb_n_s64(...) __builtin_sve_svqdecb_n_s64(__VA_ARGS__) -#define svqdecb_n_u32(...) __builtin_sve_svqdecb_n_u32(__VA_ARGS__) -#define svqdecb_n_u64(...) __builtin_sve_svqdecb_n_u64(__VA_ARGS__) -#define svqdecb_pat_n_s32(...) __builtin_sve_svqdecb_pat_n_s32(__VA_ARGS__) -#define svqdecb_pat_n_s64(...) __builtin_sve_svqdecb_pat_n_s64(__VA_ARGS__) -#define svqdecb_pat_n_u32(...) __builtin_sve_svqdecb_pat_n_u32(__VA_ARGS__) -#define svqdecb_pat_n_u64(...) __builtin_sve_svqdecb_pat_n_u64(__VA_ARGS__) -#define svqdecd_n_s32(...) __builtin_sve_svqdecd_n_s32(__VA_ARGS__) -#define svqdecd_n_s64(...) __builtin_sve_svqdecd_n_s64(__VA_ARGS__) -#define svqdecd_n_u32(...) __builtin_sve_svqdecd_n_u32(__VA_ARGS__) -#define svqdecd_n_u64(...) __builtin_sve_svqdecd_n_u64(__VA_ARGS__) -#define svqdecd_s64(...) __builtin_sve_svqdecd_s64(__VA_ARGS__) -#define svqdecd_u64(...) __builtin_sve_svqdecd_u64(__VA_ARGS__) -#define svqdecd_pat_n_s32(...) __builtin_sve_svqdecd_pat_n_s32(__VA_ARGS__) -#define svqdecd_pat_n_s64(...) __builtin_sve_svqdecd_pat_n_s64(__VA_ARGS__) -#define svqdecd_pat_n_u32(...) __builtin_sve_svqdecd_pat_n_u32(__VA_ARGS__) -#define svqdecd_pat_n_u64(...) __builtin_sve_svqdecd_pat_n_u64(__VA_ARGS__) -#define svqdecd_pat_s64(...) __builtin_sve_svqdecd_pat_s64(__VA_ARGS__) -#define svqdecd_pat_u64(...) __builtin_sve_svqdecd_pat_u64(__VA_ARGS__) -#define svqdech_n_s32(...) __builtin_sve_svqdech_n_s32(__VA_ARGS__) -#define svqdech_n_s64(...) __builtin_sve_svqdech_n_s64(__VA_ARGS__) -#define svqdech_n_u32(...) __builtin_sve_svqdech_n_u32(__VA_ARGS__) -#define svqdech_n_u64(...) __builtin_sve_svqdech_n_u64(__VA_ARGS__) -#define svqdech_s16(...) __builtin_sve_svqdech_s16(__VA_ARGS__) -#define svqdech_u16(...) __builtin_sve_svqdech_u16(__VA_ARGS__) -#define svqdech_pat_n_s32(...) __builtin_sve_svqdech_pat_n_s32(__VA_ARGS__) -#define svqdech_pat_n_s64(...) __builtin_sve_svqdech_pat_n_s64(__VA_ARGS__) -#define svqdech_pat_n_u32(...) __builtin_sve_svqdech_pat_n_u32(__VA_ARGS__) -#define svqdech_pat_n_u64(...) __builtin_sve_svqdech_pat_n_u64(__VA_ARGS__) -#define svqdech_pat_s16(...) __builtin_sve_svqdech_pat_s16(__VA_ARGS__) -#define svqdech_pat_u16(...) __builtin_sve_svqdech_pat_u16(__VA_ARGS__) -#define svqdecp_n_s32_b8(...) __builtin_sve_svqdecp_n_s32_b8(__VA_ARGS__) -#define svqdecp_n_s32_b32(...) __builtin_sve_svqdecp_n_s32_b32(__VA_ARGS__) -#define svqdecp_n_s32_b64(...) __builtin_sve_svqdecp_n_s32_b64(__VA_ARGS__) -#define svqdecp_n_s32_b16(...) __builtin_sve_svqdecp_n_s32_b16(__VA_ARGS__) -#define svqdecp_n_s64_b8(...) __builtin_sve_svqdecp_n_s64_b8(__VA_ARGS__) -#define svqdecp_n_s64_b32(...) __builtin_sve_svqdecp_n_s64_b32(__VA_ARGS__) -#define svqdecp_n_s64_b64(...) __builtin_sve_svqdecp_n_s64_b64(__VA_ARGS__) -#define svqdecp_n_s64_b16(...) __builtin_sve_svqdecp_n_s64_b16(__VA_ARGS__) -#define svqdecp_n_u32_b8(...) __builtin_sve_svqdecp_n_u32_b8(__VA_ARGS__) -#define svqdecp_n_u32_b32(...) __builtin_sve_svqdecp_n_u32_b32(__VA_ARGS__) -#define svqdecp_n_u32_b64(...) __builtin_sve_svqdecp_n_u32_b64(__VA_ARGS__) -#define svqdecp_n_u32_b16(...) __builtin_sve_svqdecp_n_u32_b16(__VA_ARGS__) -#define svqdecp_n_u64_b8(...) __builtin_sve_svqdecp_n_u64_b8(__VA_ARGS__) -#define svqdecp_n_u64_b32(...) __builtin_sve_svqdecp_n_u64_b32(__VA_ARGS__) -#define svqdecp_n_u64_b64(...) __builtin_sve_svqdecp_n_u64_b64(__VA_ARGS__) -#define svqdecp_n_u64_b16(...) __builtin_sve_svqdecp_n_u64_b16(__VA_ARGS__) -#define svqdecp_s32(...) __builtin_sve_svqdecp_s32(__VA_ARGS__) -#define svqdecp_s64(...) __builtin_sve_svqdecp_s64(__VA_ARGS__) -#define svqdecp_s16(...) __builtin_sve_svqdecp_s16(__VA_ARGS__) -#define svqdecp_u32(...) __builtin_sve_svqdecp_u32(__VA_ARGS__) -#define svqdecp_u64(...) __builtin_sve_svqdecp_u64(__VA_ARGS__) -#define svqdecp_u16(...) __builtin_sve_svqdecp_u16(__VA_ARGS__) -#define svqdecw_n_s32(...) __builtin_sve_svqdecw_n_s32(__VA_ARGS__) -#define svqdecw_n_s64(...) __builtin_sve_svqdecw_n_s64(__VA_ARGS__) -#define svqdecw_n_u32(...) __builtin_sve_svqdecw_n_u32(__VA_ARGS__) -#define svqdecw_n_u64(...) __builtin_sve_svqdecw_n_u64(__VA_ARGS__) -#define svqdecw_s32(...) __builtin_sve_svqdecw_s32(__VA_ARGS__) -#define svqdecw_u32(...) __builtin_sve_svqdecw_u32(__VA_ARGS__) -#define svqdecw_pat_n_s32(...) __builtin_sve_svqdecw_pat_n_s32(__VA_ARGS__) -#define svqdecw_pat_n_s64(...) __builtin_sve_svqdecw_pat_n_s64(__VA_ARGS__) -#define svqdecw_pat_n_u32(...) __builtin_sve_svqdecw_pat_n_u32(__VA_ARGS__) -#define svqdecw_pat_n_u64(...) __builtin_sve_svqdecw_pat_n_u64(__VA_ARGS__) -#define svqdecw_pat_s32(...) __builtin_sve_svqdecw_pat_s32(__VA_ARGS__) -#define svqdecw_pat_u32(...) __builtin_sve_svqdecw_pat_u32(__VA_ARGS__) -#define svqincb_n_s32(...) __builtin_sve_svqincb_n_s32(__VA_ARGS__) -#define svqincb_n_s64(...) __builtin_sve_svqincb_n_s64(__VA_ARGS__) -#define svqincb_n_u32(...) __builtin_sve_svqincb_n_u32(__VA_ARGS__) -#define svqincb_n_u64(...) __builtin_sve_svqincb_n_u64(__VA_ARGS__) -#define svqincb_pat_n_s32(...) __builtin_sve_svqincb_pat_n_s32(__VA_ARGS__) -#define svqincb_pat_n_s64(...) __builtin_sve_svqincb_pat_n_s64(__VA_ARGS__) -#define svqincb_pat_n_u32(...) __builtin_sve_svqincb_pat_n_u32(__VA_ARGS__) -#define svqincb_pat_n_u64(...) __builtin_sve_svqincb_pat_n_u64(__VA_ARGS__) -#define svqincd_n_s32(...) __builtin_sve_svqincd_n_s32(__VA_ARGS__) -#define svqincd_n_s64(...) __builtin_sve_svqincd_n_s64(__VA_ARGS__) -#define svqincd_n_u32(...) __builtin_sve_svqincd_n_u32(__VA_ARGS__) -#define svqincd_n_u64(...) __builtin_sve_svqincd_n_u64(__VA_ARGS__) -#define svqincd_s64(...) __builtin_sve_svqincd_s64(__VA_ARGS__) -#define svqincd_u64(...) __builtin_sve_svqincd_u64(__VA_ARGS__) -#define svqincd_pat_n_s32(...) __builtin_sve_svqincd_pat_n_s32(__VA_ARGS__) -#define svqincd_pat_n_s64(...) __builtin_sve_svqincd_pat_n_s64(__VA_ARGS__) -#define svqincd_pat_n_u32(...) __builtin_sve_svqincd_pat_n_u32(__VA_ARGS__) -#define svqincd_pat_n_u64(...) __builtin_sve_svqincd_pat_n_u64(__VA_ARGS__) -#define svqincd_pat_s64(...) __builtin_sve_svqincd_pat_s64(__VA_ARGS__) -#define svqincd_pat_u64(...) __builtin_sve_svqincd_pat_u64(__VA_ARGS__) -#define svqinch_n_s32(...) __builtin_sve_svqinch_n_s32(__VA_ARGS__) -#define svqinch_n_s64(...) __builtin_sve_svqinch_n_s64(__VA_ARGS__) -#define svqinch_n_u32(...) __builtin_sve_svqinch_n_u32(__VA_ARGS__) -#define svqinch_n_u64(...) __builtin_sve_svqinch_n_u64(__VA_ARGS__) -#define svqinch_s16(...) __builtin_sve_svqinch_s16(__VA_ARGS__) -#define svqinch_u16(...) __builtin_sve_svqinch_u16(__VA_ARGS__) -#define svqinch_pat_n_s32(...) __builtin_sve_svqinch_pat_n_s32(__VA_ARGS__) -#define svqinch_pat_n_s64(...) __builtin_sve_svqinch_pat_n_s64(__VA_ARGS__) -#define svqinch_pat_n_u32(...) __builtin_sve_svqinch_pat_n_u32(__VA_ARGS__) -#define svqinch_pat_n_u64(...) __builtin_sve_svqinch_pat_n_u64(__VA_ARGS__) -#define svqinch_pat_s16(...) __builtin_sve_svqinch_pat_s16(__VA_ARGS__) -#define svqinch_pat_u16(...) __builtin_sve_svqinch_pat_u16(__VA_ARGS__) -#define svqincp_n_s32_b8(...) __builtin_sve_svqincp_n_s32_b8(__VA_ARGS__) -#define svqincp_n_s32_b32(...) __builtin_sve_svqincp_n_s32_b32(__VA_ARGS__) -#define svqincp_n_s32_b64(...) __builtin_sve_svqincp_n_s32_b64(__VA_ARGS__) -#define svqincp_n_s32_b16(...) __builtin_sve_svqincp_n_s32_b16(__VA_ARGS__) -#define svqincp_n_s64_b8(...) __builtin_sve_svqincp_n_s64_b8(__VA_ARGS__) -#define svqincp_n_s64_b32(...) __builtin_sve_svqincp_n_s64_b32(__VA_ARGS__) -#define svqincp_n_s64_b64(...) __builtin_sve_svqincp_n_s64_b64(__VA_ARGS__) -#define svqincp_n_s64_b16(...) __builtin_sve_svqincp_n_s64_b16(__VA_ARGS__) -#define svqincp_n_u32_b8(...) __builtin_sve_svqincp_n_u32_b8(__VA_ARGS__) -#define svqincp_n_u32_b32(...) __builtin_sve_svqincp_n_u32_b32(__VA_ARGS__) -#define svqincp_n_u32_b64(...) __builtin_sve_svqincp_n_u32_b64(__VA_ARGS__) -#define svqincp_n_u32_b16(...) __builtin_sve_svqincp_n_u32_b16(__VA_ARGS__) -#define svqincp_n_u64_b8(...) __builtin_sve_svqincp_n_u64_b8(__VA_ARGS__) -#define svqincp_n_u64_b32(...) __builtin_sve_svqincp_n_u64_b32(__VA_ARGS__) -#define svqincp_n_u64_b64(...) __builtin_sve_svqincp_n_u64_b64(__VA_ARGS__) -#define svqincp_n_u64_b16(...) __builtin_sve_svqincp_n_u64_b16(__VA_ARGS__) -#define svqincp_s32(...) __builtin_sve_svqincp_s32(__VA_ARGS__) -#define svqincp_s64(...) __builtin_sve_svqincp_s64(__VA_ARGS__) -#define svqincp_s16(...) __builtin_sve_svqincp_s16(__VA_ARGS__) -#define svqincp_u32(...) __builtin_sve_svqincp_u32(__VA_ARGS__) -#define svqincp_u64(...) __builtin_sve_svqincp_u64(__VA_ARGS__) -#define svqincp_u16(...) __builtin_sve_svqincp_u16(__VA_ARGS__) -#define svqincw_n_s32(...) __builtin_sve_svqincw_n_s32(__VA_ARGS__) -#define svqincw_n_s64(...) __builtin_sve_svqincw_n_s64(__VA_ARGS__) -#define svqincw_n_u32(...) __builtin_sve_svqincw_n_u32(__VA_ARGS__) -#define svqincw_n_u64(...) __builtin_sve_svqincw_n_u64(__VA_ARGS__) -#define svqincw_s32(...) __builtin_sve_svqincw_s32(__VA_ARGS__) -#define svqincw_u32(...) __builtin_sve_svqincw_u32(__VA_ARGS__) -#define svqincw_pat_n_s32(...) __builtin_sve_svqincw_pat_n_s32(__VA_ARGS__) -#define svqincw_pat_n_s64(...) __builtin_sve_svqincw_pat_n_s64(__VA_ARGS__) -#define svqincw_pat_n_u32(...) __builtin_sve_svqincw_pat_n_u32(__VA_ARGS__) -#define svqincw_pat_n_u64(...) __builtin_sve_svqincw_pat_n_u64(__VA_ARGS__) -#define svqincw_pat_s32(...) __builtin_sve_svqincw_pat_s32(__VA_ARGS__) -#define svqincw_pat_u32(...) __builtin_sve_svqincw_pat_u32(__VA_ARGS__) -#define svqsub_n_s8(...) __builtin_sve_svqsub_n_s8(__VA_ARGS__) -#define svqsub_n_s32(...) __builtin_sve_svqsub_n_s32(__VA_ARGS__) -#define svqsub_n_s64(...) __builtin_sve_svqsub_n_s64(__VA_ARGS__) -#define svqsub_n_s16(...) __builtin_sve_svqsub_n_s16(__VA_ARGS__) -#define svqsub_n_u8(...) __builtin_sve_svqsub_n_u8(__VA_ARGS__) -#define svqsub_n_u32(...) __builtin_sve_svqsub_n_u32(__VA_ARGS__) -#define svqsub_n_u64(...) __builtin_sve_svqsub_n_u64(__VA_ARGS__) -#define svqsub_n_u16(...) __builtin_sve_svqsub_n_u16(__VA_ARGS__) -#define svqsub_s8(...) __builtin_sve_svqsub_s8(__VA_ARGS__) -#define svqsub_s32(...) __builtin_sve_svqsub_s32(__VA_ARGS__) -#define svqsub_s64(...) __builtin_sve_svqsub_s64(__VA_ARGS__) -#define svqsub_s16(...) __builtin_sve_svqsub_s16(__VA_ARGS__) -#define svqsub_u8(...) __builtin_sve_svqsub_u8(__VA_ARGS__) -#define svqsub_u32(...) __builtin_sve_svqsub_u32(__VA_ARGS__) -#define svqsub_u64(...) __builtin_sve_svqsub_u64(__VA_ARGS__) -#define svqsub_u16(...) __builtin_sve_svqsub_u16(__VA_ARGS__) -#define svrbit_u8_m(...) __builtin_sve_svrbit_u8_m(__VA_ARGS__) -#define svrbit_u32_m(...) __builtin_sve_svrbit_u32_m(__VA_ARGS__) -#define svrbit_u64_m(...) __builtin_sve_svrbit_u64_m(__VA_ARGS__) -#define svrbit_u16_m(...) __builtin_sve_svrbit_u16_m(__VA_ARGS__) -#define svrbit_s8_m(...) __builtin_sve_svrbit_s8_m(__VA_ARGS__) -#define svrbit_s32_m(...) __builtin_sve_svrbit_s32_m(__VA_ARGS__) -#define svrbit_s64_m(...) __builtin_sve_svrbit_s64_m(__VA_ARGS__) -#define svrbit_s16_m(...) __builtin_sve_svrbit_s16_m(__VA_ARGS__) -#define svrbit_u8_x(...) __builtin_sve_svrbit_u8_x(__VA_ARGS__) -#define svrbit_u32_x(...) __builtin_sve_svrbit_u32_x(__VA_ARGS__) -#define svrbit_u64_x(...) __builtin_sve_svrbit_u64_x(__VA_ARGS__) -#define svrbit_u16_x(...) __builtin_sve_svrbit_u16_x(__VA_ARGS__) -#define svrbit_s8_x(...) __builtin_sve_svrbit_s8_x(__VA_ARGS__) -#define svrbit_s32_x(...) __builtin_sve_svrbit_s32_x(__VA_ARGS__) -#define svrbit_s64_x(...) __builtin_sve_svrbit_s64_x(__VA_ARGS__) -#define svrbit_s16_x(...) __builtin_sve_svrbit_s16_x(__VA_ARGS__) -#define svrbit_u8_z(...) __builtin_sve_svrbit_u8_z(__VA_ARGS__) -#define svrbit_u32_z(...) __builtin_sve_svrbit_u32_z(__VA_ARGS__) -#define svrbit_u64_z(...) __builtin_sve_svrbit_u64_z(__VA_ARGS__) -#define svrbit_u16_z(...) __builtin_sve_svrbit_u16_z(__VA_ARGS__) -#define svrbit_s8_z(...) __builtin_sve_svrbit_s8_z(__VA_ARGS__) -#define svrbit_s32_z(...) __builtin_sve_svrbit_s32_z(__VA_ARGS__) -#define svrbit_s64_z(...) __builtin_sve_svrbit_s64_z(__VA_ARGS__) -#define svrbit_s16_z(...) __builtin_sve_svrbit_s16_z(__VA_ARGS__) -#define svrdffr(...) __builtin_sve_svrdffr(__VA_ARGS__) -#define svrdffr_z(...) __builtin_sve_svrdffr_z(__VA_ARGS__) -#define svrecpe_f64(...) __builtin_sve_svrecpe_f64(__VA_ARGS__) -#define svrecpe_f32(...) __builtin_sve_svrecpe_f32(__VA_ARGS__) -#define svrecpe_f16(...) __builtin_sve_svrecpe_f16(__VA_ARGS__) -#define svrecps_f64(...) __builtin_sve_svrecps_f64(__VA_ARGS__) -#define svrecps_f32(...) __builtin_sve_svrecps_f32(__VA_ARGS__) -#define svrecps_f16(...) __builtin_sve_svrecps_f16(__VA_ARGS__) -#define svrecpx_f64_m(...) __builtin_sve_svrecpx_f64_m(__VA_ARGS__) -#define svrecpx_f32_m(...) __builtin_sve_svrecpx_f32_m(__VA_ARGS__) -#define svrecpx_f16_m(...) __builtin_sve_svrecpx_f16_m(__VA_ARGS__) -#define svrecpx_f64_x(...) __builtin_sve_svrecpx_f64_x(__VA_ARGS__) -#define svrecpx_f32_x(...) __builtin_sve_svrecpx_f32_x(__VA_ARGS__) -#define svrecpx_f16_x(...) __builtin_sve_svrecpx_f16_x(__VA_ARGS__) -#define svrecpx_f64_z(...) __builtin_sve_svrecpx_f64_z(__VA_ARGS__) -#define svrecpx_f32_z(...) __builtin_sve_svrecpx_f32_z(__VA_ARGS__) -#define svrecpx_f16_z(...) __builtin_sve_svrecpx_f16_z(__VA_ARGS__) -#define svrev_u8(...) __builtin_sve_svrev_u8(__VA_ARGS__) -#define svrev_u32(...) __builtin_sve_svrev_u32(__VA_ARGS__) -#define svrev_u64(...) __builtin_sve_svrev_u64(__VA_ARGS__) -#define svrev_u16(...) __builtin_sve_svrev_u16(__VA_ARGS__) -#define svrev_s8(...) __builtin_sve_svrev_s8(__VA_ARGS__) -#define svrev_f64(...) __builtin_sve_svrev_f64(__VA_ARGS__) -#define svrev_f32(...) __builtin_sve_svrev_f32(__VA_ARGS__) -#define svrev_f16(...) __builtin_sve_svrev_f16(__VA_ARGS__) -#define svrev_s32(...) __builtin_sve_svrev_s32(__VA_ARGS__) -#define svrev_s64(...) __builtin_sve_svrev_s64(__VA_ARGS__) -#define svrev_s16(...) __builtin_sve_svrev_s16(__VA_ARGS__) -#define svrev_b8(...) __builtin_sve_svrev_b8(__VA_ARGS__) -#define svrev_b32(...) __builtin_sve_svrev_b32(__VA_ARGS__) -#define svrev_b64(...) __builtin_sve_svrev_b64(__VA_ARGS__) -#define svrev_b16(...) __builtin_sve_svrev_b16(__VA_ARGS__) -#define svrevb_u32_m(...) __builtin_sve_svrevb_u32_m(__VA_ARGS__) -#define svrevb_u64_m(...) __builtin_sve_svrevb_u64_m(__VA_ARGS__) -#define svrevb_u16_m(...) __builtin_sve_svrevb_u16_m(__VA_ARGS__) -#define svrevb_s32_m(...) __builtin_sve_svrevb_s32_m(__VA_ARGS__) -#define svrevb_s64_m(...) __builtin_sve_svrevb_s64_m(__VA_ARGS__) -#define svrevb_s16_m(...) __builtin_sve_svrevb_s16_m(__VA_ARGS__) -#define svrevb_u32_x(...) __builtin_sve_svrevb_u32_x(__VA_ARGS__) -#define svrevb_u64_x(...) __builtin_sve_svrevb_u64_x(__VA_ARGS__) -#define svrevb_u16_x(...) __builtin_sve_svrevb_u16_x(__VA_ARGS__) -#define svrevb_s32_x(...) __builtin_sve_svrevb_s32_x(__VA_ARGS__) -#define svrevb_s64_x(...) __builtin_sve_svrevb_s64_x(__VA_ARGS__) -#define svrevb_s16_x(...) __builtin_sve_svrevb_s16_x(__VA_ARGS__) -#define svrevb_u32_z(...) __builtin_sve_svrevb_u32_z(__VA_ARGS__) -#define svrevb_u64_z(...) __builtin_sve_svrevb_u64_z(__VA_ARGS__) -#define svrevb_u16_z(...) __builtin_sve_svrevb_u16_z(__VA_ARGS__) -#define svrevb_s32_z(...) __builtin_sve_svrevb_s32_z(__VA_ARGS__) -#define svrevb_s64_z(...) __builtin_sve_svrevb_s64_z(__VA_ARGS__) -#define svrevb_s16_z(...) __builtin_sve_svrevb_s16_z(__VA_ARGS__) -#define svrevh_u32_m(...) __builtin_sve_svrevh_u32_m(__VA_ARGS__) -#define svrevh_u64_m(...) __builtin_sve_svrevh_u64_m(__VA_ARGS__) -#define svrevh_s32_m(...) __builtin_sve_svrevh_s32_m(__VA_ARGS__) -#define svrevh_s64_m(...) __builtin_sve_svrevh_s64_m(__VA_ARGS__) -#define svrevh_u32_x(...) __builtin_sve_svrevh_u32_x(__VA_ARGS__) -#define svrevh_u64_x(...) __builtin_sve_svrevh_u64_x(__VA_ARGS__) -#define svrevh_s32_x(...) __builtin_sve_svrevh_s32_x(__VA_ARGS__) -#define svrevh_s64_x(...) __builtin_sve_svrevh_s64_x(__VA_ARGS__) -#define svrevh_u32_z(...) __builtin_sve_svrevh_u32_z(__VA_ARGS__) -#define svrevh_u64_z(...) __builtin_sve_svrevh_u64_z(__VA_ARGS__) -#define svrevh_s32_z(...) __builtin_sve_svrevh_s32_z(__VA_ARGS__) -#define svrevh_s64_z(...) __builtin_sve_svrevh_s64_z(__VA_ARGS__) -#define svrevw_u64_m(...) __builtin_sve_svrevw_u64_m(__VA_ARGS__) -#define svrevw_s64_m(...) __builtin_sve_svrevw_s64_m(__VA_ARGS__) -#define svrevw_u64_x(...) __builtin_sve_svrevw_u64_x(__VA_ARGS__) -#define svrevw_s64_x(...) __builtin_sve_svrevw_s64_x(__VA_ARGS__) -#define svrevw_u64_z(...) __builtin_sve_svrevw_u64_z(__VA_ARGS__) -#define svrevw_s64_z(...) __builtin_sve_svrevw_s64_z(__VA_ARGS__) -#define svrinta_f64_m(...) __builtin_sve_svrinta_f64_m(__VA_ARGS__) -#define svrinta_f32_m(...) __builtin_sve_svrinta_f32_m(__VA_ARGS__) -#define svrinta_f16_m(...) __builtin_sve_svrinta_f16_m(__VA_ARGS__) -#define svrinta_f64_x(...) __builtin_sve_svrinta_f64_x(__VA_ARGS__) -#define svrinta_f32_x(...) __builtin_sve_svrinta_f32_x(__VA_ARGS__) -#define svrinta_f16_x(...) __builtin_sve_svrinta_f16_x(__VA_ARGS__) -#define svrinta_f64_z(...) __builtin_sve_svrinta_f64_z(__VA_ARGS__) -#define svrinta_f32_z(...) __builtin_sve_svrinta_f32_z(__VA_ARGS__) -#define svrinta_f16_z(...) __builtin_sve_svrinta_f16_z(__VA_ARGS__) -#define svrinti_f64_m(...) __builtin_sve_svrinti_f64_m(__VA_ARGS__) -#define svrinti_f32_m(...) __builtin_sve_svrinti_f32_m(__VA_ARGS__) -#define svrinti_f16_m(...) __builtin_sve_svrinti_f16_m(__VA_ARGS__) -#define svrinti_f64_x(...) __builtin_sve_svrinti_f64_x(__VA_ARGS__) -#define svrinti_f32_x(...) __builtin_sve_svrinti_f32_x(__VA_ARGS__) -#define svrinti_f16_x(...) __builtin_sve_svrinti_f16_x(__VA_ARGS__) -#define svrinti_f64_z(...) __builtin_sve_svrinti_f64_z(__VA_ARGS__) -#define svrinti_f32_z(...) __builtin_sve_svrinti_f32_z(__VA_ARGS__) -#define svrinti_f16_z(...) __builtin_sve_svrinti_f16_z(__VA_ARGS__) -#define svrintm_f64_m(...) __builtin_sve_svrintm_f64_m(__VA_ARGS__) -#define svrintm_f32_m(...) __builtin_sve_svrintm_f32_m(__VA_ARGS__) -#define svrintm_f16_m(...) __builtin_sve_svrintm_f16_m(__VA_ARGS__) -#define svrintm_f64_x(...) __builtin_sve_svrintm_f64_x(__VA_ARGS__) -#define svrintm_f32_x(...) __builtin_sve_svrintm_f32_x(__VA_ARGS__) -#define svrintm_f16_x(...) __builtin_sve_svrintm_f16_x(__VA_ARGS__) -#define svrintm_f64_z(...) __builtin_sve_svrintm_f64_z(__VA_ARGS__) -#define svrintm_f32_z(...) __builtin_sve_svrintm_f32_z(__VA_ARGS__) -#define svrintm_f16_z(...) __builtin_sve_svrintm_f16_z(__VA_ARGS__) -#define svrintn_f64_m(...) __builtin_sve_svrintn_f64_m(__VA_ARGS__) -#define svrintn_f32_m(...) __builtin_sve_svrintn_f32_m(__VA_ARGS__) -#define svrintn_f16_m(...) __builtin_sve_svrintn_f16_m(__VA_ARGS__) -#define svrintn_f64_x(...) __builtin_sve_svrintn_f64_x(__VA_ARGS__) -#define svrintn_f32_x(...) __builtin_sve_svrintn_f32_x(__VA_ARGS__) -#define svrintn_f16_x(...) __builtin_sve_svrintn_f16_x(__VA_ARGS__) -#define svrintn_f64_z(...) __builtin_sve_svrintn_f64_z(__VA_ARGS__) -#define svrintn_f32_z(...) __builtin_sve_svrintn_f32_z(__VA_ARGS__) -#define svrintn_f16_z(...) __builtin_sve_svrintn_f16_z(__VA_ARGS__) -#define svrintp_f64_m(...) __builtin_sve_svrintp_f64_m(__VA_ARGS__) -#define svrintp_f32_m(...) __builtin_sve_svrintp_f32_m(__VA_ARGS__) -#define svrintp_f16_m(...) __builtin_sve_svrintp_f16_m(__VA_ARGS__) -#define svrintp_f64_x(...) __builtin_sve_svrintp_f64_x(__VA_ARGS__) -#define svrintp_f32_x(...) __builtin_sve_svrintp_f32_x(__VA_ARGS__) -#define svrintp_f16_x(...) __builtin_sve_svrintp_f16_x(__VA_ARGS__) -#define svrintp_f64_z(...) __builtin_sve_svrintp_f64_z(__VA_ARGS__) -#define svrintp_f32_z(...) __builtin_sve_svrintp_f32_z(__VA_ARGS__) -#define svrintp_f16_z(...) __builtin_sve_svrintp_f16_z(__VA_ARGS__) -#define svrintx_f64_m(...) __builtin_sve_svrintx_f64_m(__VA_ARGS__) -#define svrintx_f32_m(...) __builtin_sve_svrintx_f32_m(__VA_ARGS__) -#define svrintx_f16_m(...) __builtin_sve_svrintx_f16_m(__VA_ARGS__) -#define svrintx_f64_x(...) __builtin_sve_svrintx_f64_x(__VA_ARGS__) -#define svrintx_f32_x(...) __builtin_sve_svrintx_f32_x(__VA_ARGS__) -#define svrintx_f16_x(...) __builtin_sve_svrintx_f16_x(__VA_ARGS__) -#define svrintx_f64_z(...) __builtin_sve_svrintx_f64_z(__VA_ARGS__) -#define svrintx_f32_z(...) __builtin_sve_svrintx_f32_z(__VA_ARGS__) -#define svrintx_f16_z(...) __builtin_sve_svrintx_f16_z(__VA_ARGS__) -#define svrintz_f64_m(...) __builtin_sve_svrintz_f64_m(__VA_ARGS__) -#define svrintz_f32_m(...) __builtin_sve_svrintz_f32_m(__VA_ARGS__) -#define svrintz_f16_m(...) __builtin_sve_svrintz_f16_m(__VA_ARGS__) -#define svrintz_f64_x(...) __builtin_sve_svrintz_f64_x(__VA_ARGS__) -#define svrintz_f32_x(...) __builtin_sve_svrintz_f32_x(__VA_ARGS__) -#define svrintz_f16_x(...) __builtin_sve_svrintz_f16_x(__VA_ARGS__) -#define svrintz_f64_z(...) __builtin_sve_svrintz_f64_z(__VA_ARGS__) -#define svrintz_f32_z(...) __builtin_sve_svrintz_f32_z(__VA_ARGS__) -#define svrintz_f16_z(...) __builtin_sve_svrintz_f16_z(__VA_ARGS__) -#define svrsqrte_f64(...) __builtin_sve_svrsqrte_f64(__VA_ARGS__) -#define svrsqrte_f32(...) __builtin_sve_svrsqrte_f32(__VA_ARGS__) -#define svrsqrte_f16(...) __builtin_sve_svrsqrte_f16(__VA_ARGS__) -#define svrsqrts_f64(...) __builtin_sve_svrsqrts_f64(__VA_ARGS__) -#define svrsqrts_f32(...) __builtin_sve_svrsqrts_f32(__VA_ARGS__) -#define svrsqrts_f16(...) __builtin_sve_svrsqrts_f16(__VA_ARGS__) -#define svscale_n_f64_m(...) __builtin_sve_svscale_n_f64_m(__VA_ARGS__) -#define svscale_n_f32_m(...) __builtin_sve_svscale_n_f32_m(__VA_ARGS__) -#define svscale_n_f16_m(...) __builtin_sve_svscale_n_f16_m(__VA_ARGS__) -#define svscale_n_f64_x(...) __builtin_sve_svscale_n_f64_x(__VA_ARGS__) -#define svscale_n_f32_x(...) __builtin_sve_svscale_n_f32_x(__VA_ARGS__) -#define svscale_n_f16_x(...) __builtin_sve_svscale_n_f16_x(__VA_ARGS__) -#define svscale_n_f64_z(...) __builtin_sve_svscale_n_f64_z(__VA_ARGS__) -#define svscale_n_f32_z(...) __builtin_sve_svscale_n_f32_z(__VA_ARGS__) -#define svscale_n_f16_z(...) __builtin_sve_svscale_n_f16_z(__VA_ARGS__) -#define svscale_f64_m(...) __builtin_sve_svscale_f64_m(__VA_ARGS__) -#define svscale_f32_m(...) __builtin_sve_svscale_f32_m(__VA_ARGS__) -#define svscale_f16_m(...) __builtin_sve_svscale_f16_m(__VA_ARGS__) -#define svscale_f64_x(...) __builtin_sve_svscale_f64_x(__VA_ARGS__) -#define svscale_f32_x(...) __builtin_sve_svscale_f32_x(__VA_ARGS__) -#define svscale_f16_x(...) __builtin_sve_svscale_f16_x(__VA_ARGS__) -#define svscale_f64_z(...) __builtin_sve_svscale_f64_z(__VA_ARGS__) -#define svscale_f32_z(...) __builtin_sve_svscale_f32_z(__VA_ARGS__) -#define svscale_f16_z(...) __builtin_sve_svscale_f16_z(__VA_ARGS__) -#define svsel_b(...) __builtin_sve_svsel_b(__VA_ARGS__) -#define svsel_u8(...) __builtin_sve_svsel_u8(__VA_ARGS__) -#define svsel_u32(...) __builtin_sve_svsel_u32(__VA_ARGS__) -#define svsel_u64(...) __builtin_sve_svsel_u64(__VA_ARGS__) -#define svsel_u16(...) __builtin_sve_svsel_u16(__VA_ARGS__) -#define svsel_s8(...) __builtin_sve_svsel_s8(__VA_ARGS__) -#define svsel_f64(...) __builtin_sve_svsel_f64(__VA_ARGS__) -#define svsel_f32(...) __builtin_sve_svsel_f32(__VA_ARGS__) -#define svsel_f16(...) __builtin_sve_svsel_f16(__VA_ARGS__) -#define svsel_s32(...) __builtin_sve_svsel_s32(__VA_ARGS__) -#define svsel_s64(...) __builtin_sve_svsel_s64(__VA_ARGS__) -#define svsel_s16(...) __builtin_sve_svsel_s16(__VA_ARGS__) -#define svset2_u8(...) __builtin_sve_svset2_u8(__VA_ARGS__) -#define svset2_u32(...) __builtin_sve_svset2_u32(__VA_ARGS__) -#define svset2_u64(...) __builtin_sve_svset2_u64(__VA_ARGS__) -#define svset2_u16(...) __builtin_sve_svset2_u16(__VA_ARGS__) -#define svset2_s8(...) __builtin_sve_svset2_s8(__VA_ARGS__) -#define svset2_f64(...) __builtin_sve_svset2_f64(__VA_ARGS__) -#define svset2_f32(...) __builtin_sve_svset2_f32(__VA_ARGS__) -#define svset2_f16(...) __builtin_sve_svset2_f16(__VA_ARGS__) -#define svset2_s32(...) __builtin_sve_svset2_s32(__VA_ARGS__) -#define svset2_s64(...) __builtin_sve_svset2_s64(__VA_ARGS__) -#define svset2_s16(...) __builtin_sve_svset2_s16(__VA_ARGS__) -#define svset3_u8(...) __builtin_sve_svset3_u8(__VA_ARGS__) -#define svset3_u32(...) __builtin_sve_svset3_u32(__VA_ARGS__) -#define svset3_u64(...) __builtin_sve_svset3_u64(__VA_ARGS__) -#define svset3_u16(...) __builtin_sve_svset3_u16(__VA_ARGS__) -#define svset3_s8(...) __builtin_sve_svset3_s8(__VA_ARGS__) -#define svset3_f64(...) __builtin_sve_svset3_f64(__VA_ARGS__) -#define svset3_f32(...) __builtin_sve_svset3_f32(__VA_ARGS__) -#define svset3_f16(...) __builtin_sve_svset3_f16(__VA_ARGS__) -#define svset3_s32(...) __builtin_sve_svset3_s32(__VA_ARGS__) -#define svset3_s64(...) __builtin_sve_svset3_s64(__VA_ARGS__) -#define svset3_s16(...) __builtin_sve_svset3_s16(__VA_ARGS__) -#define svset4_u8(...) __builtin_sve_svset4_u8(__VA_ARGS__) -#define svset4_u32(...) __builtin_sve_svset4_u32(__VA_ARGS__) -#define svset4_u64(...) __builtin_sve_svset4_u64(__VA_ARGS__) -#define svset4_u16(...) __builtin_sve_svset4_u16(__VA_ARGS__) -#define svset4_s8(...) __builtin_sve_svset4_s8(__VA_ARGS__) -#define svset4_f64(...) __builtin_sve_svset4_f64(__VA_ARGS__) -#define svset4_f32(...) __builtin_sve_svset4_f32(__VA_ARGS__) -#define svset4_f16(...) __builtin_sve_svset4_f16(__VA_ARGS__) -#define svset4_s32(...) __builtin_sve_svset4_s32(__VA_ARGS__) -#define svset4_s64(...) __builtin_sve_svset4_s64(__VA_ARGS__) -#define svset4_s16(...) __builtin_sve_svset4_s16(__VA_ARGS__) -#define svsetffr(...) __builtin_sve_svsetffr(__VA_ARGS__) -#define svsplice_u8(...) __builtin_sve_svsplice_u8(__VA_ARGS__) -#define svsplice_u32(...) __builtin_sve_svsplice_u32(__VA_ARGS__) -#define svsplice_u64(...) __builtin_sve_svsplice_u64(__VA_ARGS__) -#define svsplice_u16(...) __builtin_sve_svsplice_u16(__VA_ARGS__) -#define svsplice_s8(...) __builtin_sve_svsplice_s8(__VA_ARGS__) -#define svsplice_f64(...) __builtin_sve_svsplice_f64(__VA_ARGS__) -#define svsplice_f32(...) __builtin_sve_svsplice_f32(__VA_ARGS__) -#define svsplice_f16(...) __builtin_sve_svsplice_f16(__VA_ARGS__) -#define svsplice_s32(...) __builtin_sve_svsplice_s32(__VA_ARGS__) -#define svsplice_s64(...) __builtin_sve_svsplice_s64(__VA_ARGS__) -#define svsplice_s16(...) __builtin_sve_svsplice_s16(__VA_ARGS__) -#define svsqrt_f64_m(...) __builtin_sve_svsqrt_f64_m(__VA_ARGS__) -#define svsqrt_f32_m(...) __builtin_sve_svsqrt_f32_m(__VA_ARGS__) -#define svsqrt_f16_m(...) __builtin_sve_svsqrt_f16_m(__VA_ARGS__) -#define svsqrt_f64_x(...) __builtin_sve_svsqrt_f64_x(__VA_ARGS__) -#define svsqrt_f32_x(...) __builtin_sve_svsqrt_f32_x(__VA_ARGS__) -#define svsqrt_f16_x(...) __builtin_sve_svsqrt_f16_x(__VA_ARGS__) -#define svsqrt_f64_z(...) __builtin_sve_svsqrt_f64_z(__VA_ARGS__) -#define svsqrt_f32_z(...) __builtin_sve_svsqrt_f32_z(__VA_ARGS__) -#define svsqrt_f16_z(...) __builtin_sve_svsqrt_f16_z(__VA_ARGS__) -#define svst1_u8(...) __builtin_sve_svst1_u8(__VA_ARGS__) -#define svst1_u32(...) __builtin_sve_svst1_u32(__VA_ARGS__) -#define svst1_u64(...) __builtin_sve_svst1_u64(__VA_ARGS__) -#define svst1_u16(...) __builtin_sve_svst1_u16(__VA_ARGS__) -#define svst1_s8(...) __builtin_sve_svst1_s8(__VA_ARGS__) -#define svst1_f64(...) __builtin_sve_svst1_f64(__VA_ARGS__) -#define svst1_f32(...) __builtin_sve_svst1_f32(__VA_ARGS__) -#define svst1_f16(...) __builtin_sve_svst1_f16(__VA_ARGS__) -#define svst1_s32(...) __builtin_sve_svst1_s32(__VA_ARGS__) -#define svst1_s64(...) __builtin_sve_svst1_s64(__VA_ARGS__) -#define svst1_s16(...) __builtin_sve_svst1_s16(__VA_ARGS__) -#define svst1_scatter_u32base_index_u32(...) __builtin_sve_svst1_scatter_u32base_index_u32(__VA_ARGS__) -#define svst1_scatter_u64base_index_u64(...) __builtin_sve_svst1_scatter_u64base_index_u64(__VA_ARGS__) -#define svst1_scatter_u64base_index_f64(...) __builtin_sve_svst1_scatter_u64base_index_f64(__VA_ARGS__) -#define svst1_scatter_u32base_index_f32(...) __builtin_sve_svst1_scatter_u32base_index_f32(__VA_ARGS__) -#define svst1_scatter_u32base_index_s32(...) __builtin_sve_svst1_scatter_u32base_index_s32(__VA_ARGS__) -#define svst1_scatter_u64base_index_s64(...) __builtin_sve_svst1_scatter_u64base_index_s64(__VA_ARGS__) -#define svst1_scatter_u32base_offset_u32(...) __builtin_sve_svst1_scatter_u32base_offset_u32(__VA_ARGS__) -#define svst1_scatter_u64base_offset_u64(...) __builtin_sve_svst1_scatter_u64base_offset_u64(__VA_ARGS__) -#define svst1_scatter_u64base_offset_f64(...) __builtin_sve_svst1_scatter_u64base_offset_f64(__VA_ARGS__) -#define svst1_scatter_u32base_offset_f32(...) __builtin_sve_svst1_scatter_u32base_offset_f32(__VA_ARGS__) -#define svst1_scatter_u32base_offset_s32(...) __builtin_sve_svst1_scatter_u32base_offset_s32(__VA_ARGS__) -#define svst1_scatter_u64base_offset_s64(...) __builtin_sve_svst1_scatter_u64base_offset_s64(__VA_ARGS__) -#define svst1_scatter_u32base_u32(...) __builtin_sve_svst1_scatter_u32base_u32(__VA_ARGS__) -#define svst1_scatter_u64base_u64(...) __builtin_sve_svst1_scatter_u64base_u64(__VA_ARGS__) -#define svst1_scatter_u64base_f64(...) __builtin_sve_svst1_scatter_u64base_f64(__VA_ARGS__) -#define svst1_scatter_u32base_f32(...) __builtin_sve_svst1_scatter_u32base_f32(__VA_ARGS__) -#define svst1_scatter_u32base_s32(...) __builtin_sve_svst1_scatter_u32base_s32(__VA_ARGS__) -#define svst1_scatter_u64base_s64(...) __builtin_sve_svst1_scatter_u64base_s64(__VA_ARGS__) -#define svst1_scatter_s32index_u32(...) __builtin_sve_svst1_scatter_s32index_u32(__VA_ARGS__) -#define svst1_scatter_s32index_f32(...) __builtin_sve_svst1_scatter_s32index_f32(__VA_ARGS__) -#define svst1_scatter_s32index_s32(...) __builtin_sve_svst1_scatter_s32index_s32(__VA_ARGS__) -#define svst1_scatter_u32index_u32(...) __builtin_sve_svst1_scatter_u32index_u32(__VA_ARGS__) -#define svst1_scatter_u32index_f32(...) __builtin_sve_svst1_scatter_u32index_f32(__VA_ARGS__) -#define svst1_scatter_u32index_s32(...) __builtin_sve_svst1_scatter_u32index_s32(__VA_ARGS__) -#define svst1_scatter_s64index_u64(...) __builtin_sve_svst1_scatter_s64index_u64(__VA_ARGS__) -#define svst1_scatter_s64index_f64(...) __builtin_sve_svst1_scatter_s64index_f64(__VA_ARGS__) -#define svst1_scatter_s64index_s64(...) __builtin_sve_svst1_scatter_s64index_s64(__VA_ARGS__) -#define svst1_scatter_u64index_u64(...) __builtin_sve_svst1_scatter_u64index_u64(__VA_ARGS__) -#define svst1_scatter_u64index_f64(...) __builtin_sve_svst1_scatter_u64index_f64(__VA_ARGS__) -#define svst1_scatter_u64index_s64(...) __builtin_sve_svst1_scatter_u64index_s64(__VA_ARGS__) -#define svst1_scatter_s32offset_u32(...) __builtin_sve_svst1_scatter_s32offset_u32(__VA_ARGS__) -#define svst1_scatter_s32offset_f32(...) __builtin_sve_svst1_scatter_s32offset_f32(__VA_ARGS__) -#define svst1_scatter_s32offset_s32(...) __builtin_sve_svst1_scatter_s32offset_s32(__VA_ARGS__) -#define svst1_scatter_u32offset_u32(...) __builtin_sve_svst1_scatter_u32offset_u32(__VA_ARGS__) -#define svst1_scatter_u32offset_f32(...) __builtin_sve_svst1_scatter_u32offset_f32(__VA_ARGS__) -#define svst1_scatter_u32offset_s32(...) __builtin_sve_svst1_scatter_u32offset_s32(__VA_ARGS__) -#define svst1_scatter_s64offset_u64(...) __builtin_sve_svst1_scatter_s64offset_u64(__VA_ARGS__) -#define svst1_scatter_s64offset_f64(...) __builtin_sve_svst1_scatter_s64offset_f64(__VA_ARGS__) -#define svst1_scatter_s64offset_s64(...) __builtin_sve_svst1_scatter_s64offset_s64(__VA_ARGS__) -#define svst1_scatter_u64offset_u64(...) __builtin_sve_svst1_scatter_u64offset_u64(__VA_ARGS__) -#define svst1_scatter_u64offset_f64(...) __builtin_sve_svst1_scatter_u64offset_f64(__VA_ARGS__) -#define svst1_scatter_u64offset_s64(...) __builtin_sve_svst1_scatter_u64offset_s64(__VA_ARGS__) -#define svst1_vnum_u8(...) __builtin_sve_svst1_vnum_u8(__VA_ARGS__) -#define svst1_vnum_u32(...) __builtin_sve_svst1_vnum_u32(__VA_ARGS__) -#define svst1_vnum_u64(...) __builtin_sve_svst1_vnum_u64(__VA_ARGS__) -#define svst1_vnum_u16(...) __builtin_sve_svst1_vnum_u16(__VA_ARGS__) -#define svst1_vnum_s8(...) __builtin_sve_svst1_vnum_s8(__VA_ARGS__) -#define svst1_vnum_f64(...) __builtin_sve_svst1_vnum_f64(__VA_ARGS__) -#define svst1_vnum_f32(...) __builtin_sve_svst1_vnum_f32(__VA_ARGS__) -#define svst1_vnum_f16(...) __builtin_sve_svst1_vnum_f16(__VA_ARGS__) -#define svst1_vnum_s32(...) __builtin_sve_svst1_vnum_s32(__VA_ARGS__) -#define svst1_vnum_s64(...) __builtin_sve_svst1_vnum_s64(__VA_ARGS__) -#define svst1_vnum_s16(...) __builtin_sve_svst1_vnum_s16(__VA_ARGS__) -#define svst1b_s32(...) __builtin_sve_svst1b_s32(__VA_ARGS__) -#define svst1b_s64(...) __builtin_sve_svst1b_s64(__VA_ARGS__) -#define svst1b_s16(...) __builtin_sve_svst1b_s16(__VA_ARGS__) -#define svst1b_u32(...) __builtin_sve_svst1b_u32(__VA_ARGS__) -#define svst1b_u64(...) __builtin_sve_svst1b_u64(__VA_ARGS__) -#define svst1b_u16(...) __builtin_sve_svst1b_u16(__VA_ARGS__) -#define svst1b_scatter_u32base_offset_u32(...) __builtin_sve_svst1b_scatter_u32base_offset_u32(__VA_ARGS__) -#define svst1b_scatter_u64base_offset_u64(...) __builtin_sve_svst1b_scatter_u64base_offset_u64(__VA_ARGS__) -#define svst1b_scatter_u32base_offset_s32(...) __builtin_sve_svst1b_scatter_u32base_offset_s32(__VA_ARGS__) -#define svst1b_scatter_u64base_offset_s64(...) __builtin_sve_svst1b_scatter_u64base_offset_s64(__VA_ARGS__) -#define svst1b_scatter_u32base_u32(...) __builtin_sve_svst1b_scatter_u32base_u32(__VA_ARGS__) -#define svst1b_scatter_u64base_u64(...) __builtin_sve_svst1b_scatter_u64base_u64(__VA_ARGS__) -#define svst1b_scatter_u32base_s32(...) __builtin_sve_svst1b_scatter_u32base_s32(__VA_ARGS__) -#define svst1b_scatter_u64base_s64(...) __builtin_sve_svst1b_scatter_u64base_s64(__VA_ARGS__) -#define svst1b_scatter_s32offset_s32(...) __builtin_sve_svst1b_scatter_s32offset_s32(__VA_ARGS__) -#define svst1b_scatter_s32offset_u32(...) __builtin_sve_svst1b_scatter_s32offset_u32(__VA_ARGS__) -#define svst1b_scatter_u32offset_s32(...) __builtin_sve_svst1b_scatter_u32offset_s32(__VA_ARGS__) -#define svst1b_scatter_u32offset_u32(...) __builtin_sve_svst1b_scatter_u32offset_u32(__VA_ARGS__) -#define svst1b_scatter_s64offset_s64(...) __builtin_sve_svst1b_scatter_s64offset_s64(__VA_ARGS__) -#define svst1b_scatter_s64offset_u64(...) __builtin_sve_svst1b_scatter_s64offset_u64(__VA_ARGS__) -#define svst1b_scatter_u64offset_s64(...) __builtin_sve_svst1b_scatter_u64offset_s64(__VA_ARGS__) -#define svst1b_scatter_u64offset_u64(...) __builtin_sve_svst1b_scatter_u64offset_u64(__VA_ARGS__) -#define svst1b_vnum_s32(...) __builtin_sve_svst1b_vnum_s32(__VA_ARGS__) -#define svst1b_vnum_s64(...) __builtin_sve_svst1b_vnum_s64(__VA_ARGS__) -#define svst1b_vnum_s16(...) __builtin_sve_svst1b_vnum_s16(__VA_ARGS__) -#define svst1b_vnum_u32(...) __builtin_sve_svst1b_vnum_u32(__VA_ARGS__) -#define svst1b_vnum_u64(...) __builtin_sve_svst1b_vnum_u64(__VA_ARGS__) -#define svst1b_vnum_u16(...) __builtin_sve_svst1b_vnum_u16(__VA_ARGS__) -#define svst1h_s32(...) __builtin_sve_svst1h_s32(__VA_ARGS__) -#define svst1h_s64(...) __builtin_sve_svst1h_s64(__VA_ARGS__) -#define svst1h_u32(...) __builtin_sve_svst1h_u32(__VA_ARGS__) -#define svst1h_u64(...) __builtin_sve_svst1h_u64(__VA_ARGS__) -#define svst1h_scatter_u32base_index_u32(...) __builtin_sve_svst1h_scatter_u32base_index_u32(__VA_ARGS__) -#define svst1h_scatter_u64base_index_u64(...) __builtin_sve_svst1h_scatter_u64base_index_u64(__VA_ARGS__) -#define svst1h_scatter_u32base_index_s32(...) __builtin_sve_svst1h_scatter_u32base_index_s32(__VA_ARGS__) -#define svst1h_scatter_u64base_index_s64(...) __builtin_sve_svst1h_scatter_u64base_index_s64(__VA_ARGS__) -#define svst1h_scatter_u32base_offset_u32(...) __builtin_sve_svst1h_scatter_u32base_offset_u32(__VA_ARGS__) -#define svst1h_scatter_u64base_offset_u64(...) __builtin_sve_svst1h_scatter_u64base_offset_u64(__VA_ARGS__) -#define svst1h_scatter_u32base_offset_s32(...) __builtin_sve_svst1h_scatter_u32base_offset_s32(__VA_ARGS__) -#define svst1h_scatter_u64base_offset_s64(...) __builtin_sve_svst1h_scatter_u64base_offset_s64(__VA_ARGS__) -#define svst1h_scatter_u32base_u32(...) __builtin_sve_svst1h_scatter_u32base_u32(__VA_ARGS__) -#define svst1h_scatter_u64base_u64(...) __builtin_sve_svst1h_scatter_u64base_u64(__VA_ARGS__) -#define svst1h_scatter_u32base_s32(...) __builtin_sve_svst1h_scatter_u32base_s32(__VA_ARGS__) -#define svst1h_scatter_u64base_s64(...) __builtin_sve_svst1h_scatter_u64base_s64(__VA_ARGS__) -#define svst1h_scatter_s32index_s32(...) __builtin_sve_svst1h_scatter_s32index_s32(__VA_ARGS__) -#define svst1h_scatter_s32index_u32(...) __builtin_sve_svst1h_scatter_s32index_u32(__VA_ARGS__) -#define svst1h_scatter_u32index_s32(...) __builtin_sve_svst1h_scatter_u32index_s32(__VA_ARGS__) -#define svst1h_scatter_u32index_u32(...) __builtin_sve_svst1h_scatter_u32index_u32(__VA_ARGS__) -#define svst1h_scatter_s64index_s64(...) __builtin_sve_svst1h_scatter_s64index_s64(__VA_ARGS__) -#define svst1h_scatter_s64index_u64(...) __builtin_sve_svst1h_scatter_s64index_u64(__VA_ARGS__) -#define svst1h_scatter_u64index_s64(...) __builtin_sve_svst1h_scatter_u64index_s64(__VA_ARGS__) -#define svst1h_scatter_u64index_u64(...) __builtin_sve_svst1h_scatter_u64index_u64(__VA_ARGS__) -#define svst1h_scatter_s32offset_s32(...) __builtin_sve_svst1h_scatter_s32offset_s32(__VA_ARGS__) -#define svst1h_scatter_s32offset_u32(...) __builtin_sve_svst1h_scatter_s32offset_u32(__VA_ARGS__) -#define svst1h_scatter_u32offset_s32(...) __builtin_sve_svst1h_scatter_u32offset_s32(__VA_ARGS__) -#define svst1h_scatter_u32offset_u32(...) __builtin_sve_svst1h_scatter_u32offset_u32(__VA_ARGS__) -#define svst1h_scatter_s64offset_s64(...) __builtin_sve_svst1h_scatter_s64offset_s64(__VA_ARGS__) -#define svst1h_scatter_s64offset_u64(...) __builtin_sve_svst1h_scatter_s64offset_u64(__VA_ARGS__) -#define svst1h_scatter_u64offset_s64(...) __builtin_sve_svst1h_scatter_u64offset_s64(__VA_ARGS__) -#define svst1h_scatter_u64offset_u64(...) __builtin_sve_svst1h_scatter_u64offset_u64(__VA_ARGS__) -#define svst1h_vnum_s32(...) __builtin_sve_svst1h_vnum_s32(__VA_ARGS__) -#define svst1h_vnum_s64(...) __builtin_sve_svst1h_vnum_s64(__VA_ARGS__) -#define svst1h_vnum_u32(...) __builtin_sve_svst1h_vnum_u32(__VA_ARGS__) -#define svst1h_vnum_u64(...) __builtin_sve_svst1h_vnum_u64(__VA_ARGS__) -#define svst1w_s64(...) __builtin_sve_svst1w_s64(__VA_ARGS__) -#define svst1w_u64(...) __builtin_sve_svst1w_u64(__VA_ARGS__) -#define svst1w_scatter_u64base_index_u64(...) __builtin_sve_svst1w_scatter_u64base_index_u64(__VA_ARGS__) -#define svst1w_scatter_u64base_index_s64(...) __builtin_sve_svst1w_scatter_u64base_index_s64(__VA_ARGS__) -#define svst1w_scatter_u64base_offset_u64(...) __builtin_sve_svst1w_scatter_u64base_offset_u64(__VA_ARGS__) -#define svst1w_scatter_u64base_offset_s64(...) __builtin_sve_svst1w_scatter_u64base_offset_s64(__VA_ARGS__) -#define svst1w_scatter_u64base_u64(...) __builtin_sve_svst1w_scatter_u64base_u64(__VA_ARGS__) -#define svst1w_scatter_u64base_s64(...) __builtin_sve_svst1w_scatter_u64base_s64(__VA_ARGS__) -#define svst1w_scatter_s64index_s64(...) __builtin_sve_svst1w_scatter_s64index_s64(__VA_ARGS__) -#define svst1w_scatter_s64index_u64(...) __builtin_sve_svst1w_scatter_s64index_u64(__VA_ARGS__) -#define svst1w_scatter_u64index_s64(...) __builtin_sve_svst1w_scatter_u64index_s64(__VA_ARGS__) -#define svst1w_scatter_u64index_u64(...) __builtin_sve_svst1w_scatter_u64index_u64(__VA_ARGS__) -#define svst1w_scatter_s64offset_s64(...) __builtin_sve_svst1w_scatter_s64offset_s64(__VA_ARGS__) -#define svst1w_scatter_s64offset_u64(...) __builtin_sve_svst1w_scatter_s64offset_u64(__VA_ARGS__) -#define svst1w_scatter_u64offset_s64(...) __builtin_sve_svst1w_scatter_u64offset_s64(__VA_ARGS__) -#define svst1w_scatter_u64offset_u64(...) __builtin_sve_svst1w_scatter_u64offset_u64(__VA_ARGS__) -#define svst1w_vnum_s64(...) __builtin_sve_svst1w_vnum_s64(__VA_ARGS__) -#define svst1w_vnum_u64(...) __builtin_sve_svst1w_vnum_u64(__VA_ARGS__) -#define svst2_u8(...) __builtin_sve_svst2_u8(__VA_ARGS__) -#define svst2_u32(...) __builtin_sve_svst2_u32(__VA_ARGS__) -#define svst2_u64(...) __builtin_sve_svst2_u64(__VA_ARGS__) -#define svst2_u16(...) __builtin_sve_svst2_u16(__VA_ARGS__) -#define svst2_s8(...) __builtin_sve_svst2_s8(__VA_ARGS__) -#define svst2_f64(...) __builtin_sve_svst2_f64(__VA_ARGS__) -#define svst2_f32(...) __builtin_sve_svst2_f32(__VA_ARGS__) -#define svst2_f16(...) __builtin_sve_svst2_f16(__VA_ARGS__) -#define svst2_s32(...) __builtin_sve_svst2_s32(__VA_ARGS__) -#define svst2_s64(...) __builtin_sve_svst2_s64(__VA_ARGS__) -#define svst2_s16(...) __builtin_sve_svst2_s16(__VA_ARGS__) -#define svst2_vnum_u8(...) __builtin_sve_svst2_vnum_u8(__VA_ARGS__) -#define svst2_vnum_u32(...) __builtin_sve_svst2_vnum_u32(__VA_ARGS__) -#define svst2_vnum_u64(...) __builtin_sve_svst2_vnum_u64(__VA_ARGS__) -#define svst2_vnum_u16(...) __builtin_sve_svst2_vnum_u16(__VA_ARGS__) -#define svst2_vnum_s8(...) __builtin_sve_svst2_vnum_s8(__VA_ARGS__) -#define svst2_vnum_f64(...) __builtin_sve_svst2_vnum_f64(__VA_ARGS__) -#define svst2_vnum_f32(...) __builtin_sve_svst2_vnum_f32(__VA_ARGS__) -#define svst2_vnum_f16(...) __builtin_sve_svst2_vnum_f16(__VA_ARGS__) -#define svst2_vnum_s32(...) __builtin_sve_svst2_vnum_s32(__VA_ARGS__) -#define svst2_vnum_s64(...) __builtin_sve_svst2_vnum_s64(__VA_ARGS__) -#define svst2_vnum_s16(...) __builtin_sve_svst2_vnum_s16(__VA_ARGS__) -#define svst3_u8(...) __builtin_sve_svst3_u8(__VA_ARGS__) -#define svst3_u32(...) __builtin_sve_svst3_u32(__VA_ARGS__) -#define svst3_u64(...) __builtin_sve_svst3_u64(__VA_ARGS__) -#define svst3_u16(...) __builtin_sve_svst3_u16(__VA_ARGS__) -#define svst3_s8(...) __builtin_sve_svst3_s8(__VA_ARGS__) -#define svst3_f64(...) __builtin_sve_svst3_f64(__VA_ARGS__) -#define svst3_f32(...) __builtin_sve_svst3_f32(__VA_ARGS__) -#define svst3_f16(...) __builtin_sve_svst3_f16(__VA_ARGS__) -#define svst3_s32(...) __builtin_sve_svst3_s32(__VA_ARGS__) -#define svst3_s64(...) __builtin_sve_svst3_s64(__VA_ARGS__) -#define svst3_s16(...) __builtin_sve_svst3_s16(__VA_ARGS__) -#define svst3_vnum_u8(...) __builtin_sve_svst3_vnum_u8(__VA_ARGS__) -#define svst3_vnum_u32(...) __builtin_sve_svst3_vnum_u32(__VA_ARGS__) -#define svst3_vnum_u64(...) __builtin_sve_svst3_vnum_u64(__VA_ARGS__) -#define svst3_vnum_u16(...) __builtin_sve_svst3_vnum_u16(__VA_ARGS__) -#define svst3_vnum_s8(...) __builtin_sve_svst3_vnum_s8(__VA_ARGS__) -#define svst3_vnum_f64(...) __builtin_sve_svst3_vnum_f64(__VA_ARGS__) -#define svst3_vnum_f32(...) __builtin_sve_svst3_vnum_f32(__VA_ARGS__) -#define svst3_vnum_f16(...) __builtin_sve_svst3_vnum_f16(__VA_ARGS__) -#define svst3_vnum_s32(...) __builtin_sve_svst3_vnum_s32(__VA_ARGS__) -#define svst3_vnum_s64(...) __builtin_sve_svst3_vnum_s64(__VA_ARGS__) -#define svst3_vnum_s16(...) __builtin_sve_svst3_vnum_s16(__VA_ARGS__) -#define svst4_u8(...) __builtin_sve_svst4_u8(__VA_ARGS__) -#define svst4_u32(...) __builtin_sve_svst4_u32(__VA_ARGS__) -#define svst4_u64(...) __builtin_sve_svst4_u64(__VA_ARGS__) -#define svst4_u16(...) __builtin_sve_svst4_u16(__VA_ARGS__) -#define svst4_s8(...) __builtin_sve_svst4_s8(__VA_ARGS__) -#define svst4_f64(...) __builtin_sve_svst4_f64(__VA_ARGS__) -#define svst4_f32(...) __builtin_sve_svst4_f32(__VA_ARGS__) -#define svst4_f16(...) __builtin_sve_svst4_f16(__VA_ARGS__) -#define svst4_s32(...) __builtin_sve_svst4_s32(__VA_ARGS__) -#define svst4_s64(...) __builtin_sve_svst4_s64(__VA_ARGS__) -#define svst4_s16(...) __builtin_sve_svst4_s16(__VA_ARGS__) -#define svst4_vnum_u8(...) __builtin_sve_svst4_vnum_u8(__VA_ARGS__) -#define svst4_vnum_u32(...) __builtin_sve_svst4_vnum_u32(__VA_ARGS__) -#define svst4_vnum_u64(...) __builtin_sve_svst4_vnum_u64(__VA_ARGS__) -#define svst4_vnum_u16(...) __builtin_sve_svst4_vnum_u16(__VA_ARGS__) -#define svst4_vnum_s8(...) __builtin_sve_svst4_vnum_s8(__VA_ARGS__) -#define svst4_vnum_f64(...) __builtin_sve_svst4_vnum_f64(__VA_ARGS__) -#define svst4_vnum_f32(...) __builtin_sve_svst4_vnum_f32(__VA_ARGS__) -#define svst4_vnum_f16(...) __builtin_sve_svst4_vnum_f16(__VA_ARGS__) -#define svst4_vnum_s32(...) __builtin_sve_svst4_vnum_s32(__VA_ARGS__) -#define svst4_vnum_s64(...) __builtin_sve_svst4_vnum_s64(__VA_ARGS__) -#define svst4_vnum_s16(...) __builtin_sve_svst4_vnum_s16(__VA_ARGS__) -#define svstnt1_u8(...) __builtin_sve_svstnt1_u8(__VA_ARGS__) -#define svstnt1_u32(...) __builtin_sve_svstnt1_u32(__VA_ARGS__) -#define svstnt1_u64(...) __builtin_sve_svstnt1_u64(__VA_ARGS__) -#define svstnt1_u16(...) __builtin_sve_svstnt1_u16(__VA_ARGS__) -#define svstnt1_s8(...) __builtin_sve_svstnt1_s8(__VA_ARGS__) -#define svstnt1_f64(...) __builtin_sve_svstnt1_f64(__VA_ARGS__) -#define svstnt1_f32(...) __builtin_sve_svstnt1_f32(__VA_ARGS__) -#define svstnt1_f16(...) __builtin_sve_svstnt1_f16(__VA_ARGS__) -#define svstnt1_s32(...) __builtin_sve_svstnt1_s32(__VA_ARGS__) -#define svstnt1_s64(...) __builtin_sve_svstnt1_s64(__VA_ARGS__) -#define svstnt1_s16(...) __builtin_sve_svstnt1_s16(__VA_ARGS__) -#define svstnt1_vnum_u8(...) __builtin_sve_svstnt1_vnum_u8(__VA_ARGS__) -#define svstnt1_vnum_u32(...) __builtin_sve_svstnt1_vnum_u32(__VA_ARGS__) -#define svstnt1_vnum_u64(...) __builtin_sve_svstnt1_vnum_u64(__VA_ARGS__) -#define svstnt1_vnum_u16(...) __builtin_sve_svstnt1_vnum_u16(__VA_ARGS__) -#define svstnt1_vnum_s8(...) __builtin_sve_svstnt1_vnum_s8(__VA_ARGS__) -#define svstnt1_vnum_f64(...) __builtin_sve_svstnt1_vnum_f64(__VA_ARGS__) -#define svstnt1_vnum_f32(...) __builtin_sve_svstnt1_vnum_f32(__VA_ARGS__) -#define svstnt1_vnum_f16(...) __builtin_sve_svstnt1_vnum_f16(__VA_ARGS__) -#define svstnt1_vnum_s32(...) __builtin_sve_svstnt1_vnum_s32(__VA_ARGS__) -#define svstnt1_vnum_s64(...) __builtin_sve_svstnt1_vnum_s64(__VA_ARGS__) -#define svstnt1_vnum_s16(...) __builtin_sve_svstnt1_vnum_s16(__VA_ARGS__) -#define svsub_n_f64_m(...) __builtin_sve_svsub_n_f64_m(__VA_ARGS__) -#define svsub_n_f32_m(...) __builtin_sve_svsub_n_f32_m(__VA_ARGS__) -#define svsub_n_f16_m(...) __builtin_sve_svsub_n_f16_m(__VA_ARGS__) -#define svsub_n_f64_x(...) __builtin_sve_svsub_n_f64_x(__VA_ARGS__) -#define svsub_n_f32_x(...) __builtin_sve_svsub_n_f32_x(__VA_ARGS__) -#define svsub_n_f16_x(...) __builtin_sve_svsub_n_f16_x(__VA_ARGS__) -#define svsub_n_f64_z(...) __builtin_sve_svsub_n_f64_z(__VA_ARGS__) -#define svsub_n_f32_z(...) __builtin_sve_svsub_n_f32_z(__VA_ARGS__) -#define svsub_n_f16_z(...) __builtin_sve_svsub_n_f16_z(__VA_ARGS__) -#define svsub_n_u8_m(...) __builtin_sve_svsub_n_u8_m(__VA_ARGS__) -#define svsub_n_u32_m(...) __builtin_sve_svsub_n_u32_m(__VA_ARGS__) -#define svsub_n_u64_m(...) __builtin_sve_svsub_n_u64_m(__VA_ARGS__) -#define svsub_n_u16_m(...) __builtin_sve_svsub_n_u16_m(__VA_ARGS__) -#define svsub_n_s8_m(...) __builtin_sve_svsub_n_s8_m(__VA_ARGS__) -#define svsub_n_s32_m(...) __builtin_sve_svsub_n_s32_m(__VA_ARGS__) -#define svsub_n_s64_m(...) __builtin_sve_svsub_n_s64_m(__VA_ARGS__) -#define svsub_n_s16_m(...) __builtin_sve_svsub_n_s16_m(__VA_ARGS__) -#define svsub_n_u8_x(...) __builtin_sve_svsub_n_u8_x(__VA_ARGS__) -#define svsub_n_u32_x(...) __builtin_sve_svsub_n_u32_x(__VA_ARGS__) -#define svsub_n_u64_x(...) __builtin_sve_svsub_n_u64_x(__VA_ARGS__) -#define svsub_n_u16_x(...) __builtin_sve_svsub_n_u16_x(__VA_ARGS__) -#define svsub_n_s8_x(...) __builtin_sve_svsub_n_s8_x(__VA_ARGS__) -#define svsub_n_s32_x(...) __builtin_sve_svsub_n_s32_x(__VA_ARGS__) -#define svsub_n_s64_x(...) __builtin_sve_svsub_n_s64_x(__VA_ARGS__) -#define svsub_n_s16_x(...) __builtin_sve_svsub_n_s16_x(__VA_ARGS__) -#define svsub_n_u8_z(...) __builtin_sve_svsub_n_u8_z(__VA_ARGS__) -#define svsub_n_u32_z(...) __builtin_sve_svsub_n_u32_z(__VA_ARGS__) -#define svsub_n_u64_z(...) __builtin_sve_svsub_n_u64_z(__VA_ARGS__) -#define svsub_n_u16_z(...) __builtin_sve_svsub_n_u16_z(__VA_ARGS__) -#define svsub_n_s8_z(...) __builtin_sve_svsub_n_s8_z(__VA_ARGS__) -#define svsub_n_s32_z(...) __builtin_sve_svsub_n_s32_z(__VA_ARGS__) -#define svsub_n_s64_z(...) __builtin_sve_svsub_n_s64_z(__VA_ARGS__) -#define svsub_n_s16_z(...) __builtin_sve_svsub_n_s16_z(__VA_ARGS__) -#define svsub_f64_m(...) __builtin_sve_svsub_f64_m(__VA_ARGS__) -#define svsub_f32_m(...) __builtin_sve_svsub_f32_m(__VA_ARGS__) -#define svsub_f16_m(...) __builtin_sve_svsub_f16_m(__VA_ARGS__) -#define svsub_f64_x(...) __builtin_sve_svsub_f64_x(__VA_ARGS__) -#define svsub_f32_x(...) __builtin_sve_svsub_f32_x(__VA_ARGS__) -#define svsub_f16_x(...) __builtin_sve_svsub_f16_x(__VA_ARGS__) -#define svsub_f64_z(...) __builtin_sve_svsub_f64_z(__VA_ARGS__) -#define svsub_f32_z(...) __builtin_sve_svsub_f32_z(__VA_ARGS__) -#define svsub_f16_z(...) __builtin_sve_svsub_f16_z(__VA_ARGS__) -#define svsub_u8_m(...) __builtin_sve_svsub_u8_m(__VA_ARGS__) -#define svsub_u32_m(...) __builtin_sve_svsub_u32_m(__VA_ARGS__) -#define svsub_u64_m(...) __builtin_sve_svsub_u64_m(__VA_ARGS__) -#define svsub_u16_m(...) __builtin_sve_svsub_u16_m(__VA_ARGS__) -#define svsub_s8_m(...) __builtin_sve_svsub_s8_m(__VA_ARGS__) -#define svsub_s32_m(...) __builtin_sve_svsub_s32_m(__VA_ARGS__) -#define svsub_s64_m(...) __builtin_sve_svsub_s64_m(__VA_ARGS__) -#define svsub_s16_m(...) __builtin_sve_svsub_s16_m(__VA_ARGS__) -#define svsub_u8_x(...) __builtin_sve_svsub_u8_x(__VA_ARGS__) -#define svsub_u32_x(...) __builtin_sve_svsub_u32_x(__VA_ARGS__) -#define svsub_u64_x(...) __builtin_sve_svsub_u64_x(__VA_ARGS__) -#define svsub_u16_x(...) __builtin_sve_svsub_u16_x(__VA_ARGS__) -#define svsub_s8_x(...) __builtin_sve_svsub_s8_x(__VA_ARGS__) -#define svsub_s32_x(...) __builtin_sve_svsub_s32_x(__VA_ARGS__) -#define svsub_s64_x(...) __builtin_sve_svsub_s64_x(__VA_ARGS__) -#define svsub_s16_x(...) __builtin_sve_svsub_s16_x(__VA_ARGS__) -#define svsub_u8_z(...) __builtin_sve_svsub_u8_z(__VA_ARGS__) -#define svsub_u32_z(...) __builtin_sve_svsub_u32_z(__VA_ARGS__) -#define svsub_u64_z(...) __builtin_sve_svsub_u64_z(__VA_ARGS__) -#define svsub_u16_z(...) __builtin_sve_svsub_u16_z(__VA_ARGS__) -#define svsub_s8_z(...) __builtin_sve_svsub_s8_z(__VA_ARGS__) -#define svsub_s32_z(...) __builtin_sve_svsub_s32_z(__VA_ARGS__) -#define svsub_s64_z(...) __builtin_sve_svsub_s64_z(__VA_ARGS__) -#define svsub_s16_z(...) __builtin_sve_svsub_s16_z(__VA_ARGS__) -#define svsubr_n_f64_m(...) __builtin_sve_svsubr_n_f64_m(__VA_ARGS__) -#define svsubr_n_f32_m(...) __builtin_sve_svsubr_n_f32_m(__VA_ARGS__) -#define svsubr_n_f16_m(...) __builtin_sve_svsubr_n_f16_m(__VA_ARGS__) -#define svsubr_n_f64_x(...) __builtin_sve_svsubr_n_f64_x(__VA_ARGS__) -#define svsubr_n_f32_x(...) __builtin_sve_svsubr_n_f32_x(__VA_ARGS__) -#define svsubr_n_f16_x(...) __builtin_sve_svsubr_n_f16_x(__VA_ARGS__) -#define svsubr_n_f64_z(...) __builtin_sve_svsubr_n_f64_z(__VA_ARGS__) -#define svsubr_n_f32_z(...) __builtin_sve_svsubr_n_f32_z(__VA_ARGS__) -#define svsubr_n_f16_z(...) __builtin_sve_svsubr_n_f16_z(__VA_ARGS__) -#define svsubr_n_u8_m(...) __builtin_sve_svsubr_n_u8_m(__VA_ARGS__) -#define svsubr_n_u32_m(...) __builtin_sve_svsubr_n_u32_m(__VA_ARGS__) -#define svsubr_n_u64_m(...) __builtin_sve_svsubr_n_u64_m(__VA_ARGS__) -#define svsubr_n_u16_m(...) __builtin_sve_svsubr_n_u16_m(__VA_ARGS__) -#define svsubr_n_s8_m(...) __builtin_sve_svsubr_n_s8_m(__VA_ARGS__) -#define svsubr_n_s32_m(...) __builtin_sve_svsubr_n_s32_m(__VA_ARGS__) -#define svsubr_n_s64_m(...) __builtin_sve_svsubr_n_s64_m(__VA_ARGS__) -#define svsubr_n_s16_m(...) __builtin_sve_svsubr_n_s16_m(__VA_ARGS__) -#define svsubr_n_u8_x(...) __builtin_sve_svsubr_n_u8_x(__VA_ARGS__) -#define svsubr_n_u32_x(...) __builtin_sve_svsubr_n_u32_x(__VA_ARGS__) -#define svsubr_n_u64_x(...) __builtin_sve_svsubr_n_u64_x(__VA_ARGS__) -#define svsubr_n_u16_x(...) __builtin_sve_svsubr_n_u16_x(__VA_ARGS__) -#define svsubr_n_s8_x(...) __builtin_sve_svsubr_n_s8_x(__VA_ARGS__) -#define svsubr_n_s32_x(...) __builtin_sve_svsubr_n_s32_x(__VA_ARGS__) -#define svsubr_n_s64_x(...) __builtin_sve_svsubr_n_s64_x(__VA_ARGS__) -#define svsubr_n_s16_x(...) __builtin_sve_svsubr_n_s16_x(__VA_ARGS__) -#define svsubr_n_u8_z(...) __builtin_sve_svsubr_n_u8_z(__VA_ARGS__) -#define svsubr_n_u32_z(...) __builtin_sve_svsubr_n_u32_z(__VA_ARGS__) -#define svsubr_n_u64_z(...) __builtin_sve_svsubr_n_u64_z(__VA_ARGS__) -#define svsubr_n_u16_z(...) __builtin_sve_svsubr_n_u16_z(__VA_ARGS__) -#define svsubr_n_s8_z(...) __builtin_sve_svsubr_n_s8_z(__VA_ARGS__) -#define svsubr_n_s32_z(...) __builtin_sve_svsubr_n_s32_z(__VA_ARGS__) -#define svsubr_n_s64_z(...) __builtin_sve_svsubr_n_s64_z(__VA_ARGS__) -#define svsubr_n_s16_z(...) __builtin_sve_svsubr_n_s16_z(__VA_ARGS__) -#define svsubr_f64_m(...) __builtin_sve_svsubr_f64_m(__VA_ARGS__) -#define svsubr_f32_m(...) __builtin_sve_svsubr_f32_m(__VA_ARGS__) -#define svsubr_f16_m(...) __builtin_sve_svsubr_f16_m(__VA_ARGS__) -#define svsubr_f64_x(...) __builtin_sve_svsubr_f64_x(__VA_ARGS__) -#define svsubr_f32_x(...) __builtin_sve_svsubr_f32_x(__VA_ARGS__) -#define svsubr_f16_x(...) __builtin_sve_svsubr_f16_x(__VA_ARGS__) -#define svsubr_f64_z(...) __builtin_sve_svsubr_f64_z(__VA_ARGS__) -#define svsubr_f32_z(...) __builtin_sve_svsubr_f32_z(__VA_ARGS__) -#define svsubr_f16_z(...) __builtin_sve_svsubr_f16_z(__VA_ARGS__) -#define svsubr_u8_m(...) __builtin_sve_svsubr_u8_m(__VA_ARGS__) -#define svsubr_u32_m(...) __builtin_sve_svsubr_u32_m(__VA_ARGS__) -#define svsubr_u64_m(...) __builtin_sve_svsubr_u64_m(__VA_ARGS__) -#define svsubr_u16_m(...) __builtin_sve_svsubr_u16_m(__VA_ARGS__) -#define svsubr_s8_m(...) __builtin_sve_svsubr_s8_m(__VA_ARGS__) -#define svsubr_s32_m(...) __builtin_sve_svsubr_s32_m(__VA_ARGS__) -#define svsubr_s64_m(...) __builtin_sve_svsubr_s64_m(__VA_ARGS__) -#define svsubr_s16_m(...) __builtin_sve_svsubr_s16_m(__VA_ARGS__) -#define svsubr_u8_x(...) __builtin_sve_svsubr_u8_x(__VA_ARGS__) -#define svsubr_u32_x(...) __builtin_sve_svsubr_u32_x(__VA_ARGS__) -#define svsubr_u64_x(...) __builtin_sve_svsubr_u64_x(__VA_ARGS__) -#define svsubr_u16_x(...) __builtin_sve_svsubr_u16_x(__VA_ARGS__) -#define svsubr_s8_x(...) __builtin_sve_svsubr_s8_x(__VA_ARGS__) -#define svsubr_s32_x(...) __builtin_sve_svsubr_s32_x(__VA_ARGS__) -#define svsubr_s64_x(...) __builtin_sve_svsubr_s64_x(__VA_ARGS__) -#define svsubr_s16_x(...) __builtin_sve_svsubr_s16_x(__VA_ARGS__) -#define svsubr_u8_z(...) __builtin_sve_svsubr_u8_z(__VA_ARGS__) -#define svsubr_u32_z(...) __builtin_sve_svsubr_u32_z(__VA_ARGS__) -#define svsubr_u64_z(...) __builtin_sve_svsubr_u64_z(__VA_ARGS__) -#define svsubr_u16_z(...) __builtin_sve_svsubr_u16_z(__VA_ARGS__) -#define svsubr_s8_z(...) __builtin_sve_svsubr_s8_z(__VA_ARGS__) -#define svsubr_s32_z(...) __builtin_sve_svsubr_s32_z(__VA_ARGS__) -#define svsubr_s64_z(...) __builtin_sve_svsubr_s64_z(__VA_ARGS__) -#define svsubr_s16_z(...) __builtin_sve_svsubr_s16_z(__VA_ARGS__) -#define svtbl_u8(...) __builtin_sve_svtbl_u8(__VA_ARGS__) -#define svtbl_u32(...) __builtin_sve_svtbl_u32(__VA_ARGS__) -#define svtbl_u64(...) __builtin_sve_svtbl_u64(__VA_ARGS__) -#define svtbl_u16(...) __builtin_sve_svtbl_u16(__VA_ARGS__) -#define svtbl_s8(...) __builtin_sve_svtbl_s8(__VA_ARGS__) -#define svtbl_f64(...) __builtin_sve_svtbl_f64(__VA_ARGS__) -#define svtbl_f32(...) __builtin_sve_svtbl_f32(__VA_ARGS__) -#define svtbl_f16(...) __builtin_sve_svtbl_f16(__VA_ARGS__) -#define svtbl_s32(...) __builtin_sve_svtbl_s32(__VA_ARGS__) -#define svtbl_s64(...) __builtin_sve_svtbl_s64(__VA_ARGS__) -#define svtbl_s16(...) __builtin_sve_svtbl_s16(__VA_ARGS__) -#define svtmad_f64(...) __builtin_sve_svtmad_f64(__VA_ARGS__) -#define svtmad_f32(...) __builtin_sve_svtmad_f32(__VA_ARGS__) -#define svtmad_f16(...) __builtin_sve_svtmad_f16(__VA_ARGS__) -#define svtrn1_u8(...) __builtin_sve_svtrn1_u8(__VA_ARGS__) -#define svtrn1_u32(...) __builtin_sve_svtrn1_u32(__VA_ARGS__) -#define svtrn1_u64(...) __builtin_sve_svtrn1_u64(__VA_ARGS__) -#define svtrn1_u16(...) __builtin_sve_svtrn1_u16(__VA_ARGS__) -#define svtrn1_s8(...) __builtin_sve_svtrn1_s8(__VA_ARGS__) -#define svtrn1_f64(...) __builtin_sve_svtrn1_f64(__VA_ARGS__) -#define svtrn1_f32(...) __builtin_sve_svtrn1_f32(__VA_ARGS__) -#define svtrn1_f16(...) __builtin_sve_svtrn1_f16(__VA_ARGS__) -#define svtrn1_s32(...) __builtin_sve_svtrn1_s32(__VA_ARGS__) -#define svtrn1_s64(...) __builtin_sve_svtrn1_s64(__VA_ARGS__) -#define svtrn1_s16(...) __builtin_sve_svtrn1_s16(__VA_ARGS__) -#define svtrn1_b8(...) __builtin_sve_svtrn1_b8(__VA_ARGS__) -#define svtrn1_b32(...) __builtin_sve_svtrn1_b32(__VA_ARGS__) -#define svtrn1_b64(...) __builtin_sve_svtrn1_b64(__VA_ARGS__) -#define svtrn1_b16(...) __builtin_sve_svtrn1_b16(__VA_ARGS__) -#define svtrn2_u8(...) __builtin_sve_svtrn2_u8(__VA_ARGS__) -#define svtrn2_u32(...) __builtin_sve_svtrn2_u32(__VA_ARGS__) -#define svtrn2_u64(...) __builtin_sve_svtrn2_u64(__VA_ARGS__) -#define svtrn2_u16(...) __builtin_sve_svtrn2_u16(__VA_ARGS__) -#define svtrn2_s8(...) __builtin_sve_svtrn2_s8(__VA_ARGS__) -#define svtrn2_f64(...) __builtin_sve_svtrn2_f64(__VA_ARGS__) -#define svtrn2_f32(...) __builtin_sve_svtrn2_f32(__VA_ARGS__) -#define svtrn2_f16(...) __builtin_sve_svtrn2_f16(__VA_ARGS__) -#define svtrn2_s32(...) __builtin_sve_svtrn2_s32(__VA_ARGS__) -#define svtrn2_s64(...) __builtin_sve_svtrn2_s64(__VA_ARGS__) -#define svtrn2_s16(...) __builtin_sve_svtrn2_s16(__VA_ARGS__) -#define svtrn2_b8(...) __builtin_sve_svtrn2_b8(__VA_ARGS__) -#define svtrn2_b32(...) __builtin_sve_svtrn2_b32(__VA_ARGS__) -#define svtrn2_b64(...) __builtin_sve_svtrn2_b64(__VA_ARGS__) -#define svtrn2_b16(...) __builtin_sve_svtrn2_b16(__VA_ARGS__) -#define svtsmul_f64(...) __builtin_sve_svtsmul_f64(__VA_ARGS__) -#define svtsmul_f32(...) __builtin_sve_svtsmul_f32(__VA_ARGS__) -#define svtsmul_f16(...) __builtin_sve_svtsmul_f16(__VA_ARGS__) -#define svtssel_f64(...) __builtin_sve_svtssel_f64(__VA_ARGS__) -#define svtssel_f32(...) __builtin_sve_svtssel_f32(__VA_ARGS__) -#define svtssel_f16(...) __builtin_sve_svtssel_f16(__VA_ARGS__) -#define svundef2_u8(...) __builtin_sve_svundef2_u8(__VA_ARGS__) -#define svundef2_u32(...) __builtin_sve_svundef2_u32(__VA_ARGS__) -#define svundef2_u64(...) __builtin_sve_svundef2_u64(__VA_ARGS__) -#define svundef2_u16(...) __builtin_sve_svundef2_u16(__VA_ARGS__) -#define svundef2_s8(...) __builtin_sve_svundef2_s8(__VA_ARGS__) -#define svundef2_f64(...) __builtin_sve_svundef2_f64(__VA_ARGS__) -#define svundef2_f32(...) __builtin_sve_svundef2_f32(__VA_ARGS__) -#define svundef2_f16(...) __builtin_sve_svundef2_f16(__VA_ARGS__) -#define svundef2_s32(...) __builtin_sve_svundef2_s32(__VA_ARGS__) -#define svundef2_s64(...) __builtin_sve_svundef2_s64(__VA_ARGS__) -#define svundef2_s16(...) __builtin_sve_svundef2_s16(__VA_ARGS__) -#define svundef3_u8(...) __builtin_sve_svundef3_u8(__VA_ARGS__) -#define svundef3_u32(...) __builtin_sve_svundef3_u32(__VA_ARGS__) -#define svundef3_u64(...) __builtin_sve_svundef3_u64(__VA_ARGS__) -#define svundef3_u16(...) __builtin_sve_svundef3_u16(__VA_ARGS__) -#define svundef3_s8(...) __builtin_sve_svundef3_s8(__VA_ARGS__) -#define svundef3_f64(...) __builtin_sve_svundef3_f64(__VA_ARGS__) -#define svundef3_f32(...) __builtin_sve_svundef3_f32(__VA_ARGS__) -#define svundef3_f16(...) __builtin_sve_svundef3_f16(__VA_ARGS__) -#define svundef3_s32(...) __builtin_sve_svundef3_s32(__VA_ARGS__) -#define svundef3_s64(...) __builtin_sve_svundef3_s64(__VA_ARGS__) -#define svundef3_s16(...) __builtin_sve_svundef3_s16(__VA_ARGS__) -#define svundef4_u8(...) __builtin_sve_svundef4_u8(__VA_ARGS__) -#define svundef4_u32(...) __builtin_sve_svundef4_u32(__VA_ARGS__) -#define svundef4_u64(...) __builtin_sve_svundef4_u64(__VA_ARGS__) -#define svundef4_u16(...) __builtin_sve_svundef4_u16(__VA_ARGS__) -#define svundef4_s8(...) __builtin_sve_svundef4_s8(__VA_ARGS__) -#define svundef4_f64(...) __builtin_sve_svundef4_f64(__VA_ARGS__) -#define svundef4_f32(...) __builtin_sve_svundef4_f32(__VA_ARGS__) -#define svundef4_f16(...) __builtin_sve_svundef4_f16(__VA_ARGS__) -#define svundef4_s32(...) __builtin_sve_svundef4_s32(__VA_ARGS__) -#define svundef4_s64(...) __builtin_sve_svundef4_s64(__VA_ARGS__) -#define svundef4_s16(...) __builtin_sve_svundef4_s16(__VA_ARGS__) -#define svundef_u8(...) __builtin_sve_svundef_u8(__VA_ARGS__) -#define svundef_u32(...) __builtin_sve_svundef_u32(__VA_ARGS__) -#define svundef_u64(...) __builtin_sve_svundef_u64(__VA_ARGS__) -#define svundef_u16(...) __builtin_sve_svundef_u16(__VA_ARGS__) -#define svundef_s8(...) __builtin_sve_svundef_s8(__VA_ARGS__) -#define svundef_f64(...) __builtin_sve_svundef_f64(__VA_ARGS__) -#define svundef_f32(...) __builtin_sve_svundef_f32(__VA_ARGS__) -#define svundef_f16(...) __builtin_sve_svundef_f16(__VA_ARGS__) -#define svundef_s32(...) __builtin_sve_svundef_s32(__VA_ARGS__) -#define svundef_s64(...) __builtin_sve_svundef_s64(__VA_ARGS__) -#define svundef_s16(...) __builtin_sve_svundef_s16(__VA_ARGS__) -#define svunpkhi_b(...) __builtin_sve_svunpkhi_b(__VA_ARGS__) -#define svunpkhi_s32(...) __builtin_sve_svunpkhi_s32(__VA_ARGS__) -#define svunpkhi_s64(...) __builtin_sve_svunpkhi_s64(__VA_ARGS__) -#define svunpkhi_s16(...) __builtin_sve_svunpkhi_s16(__VA_ARGS__) -#define svunpkhi_u32(...) __builtin_sve_svunpkhi_u32(__VA_ARGS__) -#define svunpkhi_u64(...) __builtin_sve_svunpkhi_u64(__VA_ARGS__) -#define svunpkhi_u16(...) __builtin_sve_svunpkhi_u16(__VA_ARGS__) -#define svunpklo_b(...) __builtin_sve_svunpklo_b(__VA_ARGS__) -#define svunpklo_s32(...) __builtin_sve_svunpklo_s32(__VA_ARGS__) -#define svunpklo_s64(...) __builtin_sve_svunpklo_s64(__VA_ARGS__) -#define svunpklo_s16(...) __builtin_sve_svunpklo_s16(__VA_ARGS__) -#define svunpklo_u32(...) __builtin_sve_svunpklo_u32(__VA_ARGS__) -#define svunpklo_u64(...) __builtin_sve_svunpklo_u64(__VA_ARGS__) -#define svunpklo_u16(...) __builtin_sve_svunpklo_u16(__VA_ARGS__) -#define svuzp1_u8(...) __builtin_sve_svuzp1_u8(__VA_ARGS__) -#define svuzp1_u32(...) __builtin_sve_svuzp1_u32(__VA_ARGS__) -#define svuzp1_u64(...) __builtin_sve_svuzp1_u64(__VA_ARGS__) -#define svuzp1_u16(...) __builtin_sve_svuzp1_u16(__VA_ARGS__) -#define svuzp1_s8(...) __builtin_sve_svuzp1_s8(__VA_ARGS__) -#define svuzp1_f64(...) __builtin_sve_svuzp1_f64(__VA_ARGS__) -#define svuzp1_f32(...) __builtin_sve_svuzp1_f32(__VA_ARGS__) -#define svuzp1_f16(...) __builtin_sve_svuzp1_f16(__VA_ARGS__) -#define svuzp1_s32(...) __builtin_sve_svuzp1_s32(__VA_ARGS__) -#define svuzp1_s64(...) __builtin_sve_svuzp1_s64(__VA_ARGS__) -#define svuzp1_s16(...) __builtin_sve_svuzp1_s16(__VA_ARGS__) -#define svuzp1_b8(...) __builtin_sve_svuzp1_b8(__VA_ARGS__) -#define svuzp1_b32(...) __builtin_sve_svuzp1_b32(__VA_ARGS__) -#define svuzp1_b64(...) __builtin_sve_svuzp1_b64(__VA_ARGS__) -#define svuzp1_b16(...) __builtin_sve_svuzp1_b16(__VA_ARGS__) -#define svuzp2_u8(...) __builtin_sve_svuzp2_u8(__VA_ARGS__) -#define svuzp2_u32(...) __builtin_sve_svuzp2_u32(__VA_ARGS__) -#define svuzp2_u64(...) __builtin_sve_svuzp2_u64(__VA_ARGS__) -#define svuzp2_u16(...) __builtin_sve_svuzp2_u16(__VA_ARGS__) -#define svuzp2_s8(...) __builtin_sve_svuzp2_s8(__VA_ARGS__) -#define svuzp2_f64(...) __builtin_sve_svuzp2_f64(__VA_ARGS__) -#define svuzp2_f32(...) __builtin_sve_svuzp2_f32(__VA_ARGS__) -#define svuzp2_f16(...) __builtin_sve_svuzp2_f16(__VA_ARGS__) -#define svuzp2_s32(...) __builtin_sve_svuzp2_s32(__VA_ARGS__) -#define svuzp2_s64(...) __builtin_sve_svuzp2_s64(__VA_ARGS__) -#define svuzp2_s16(...) __builtin_sve_svuzp2_s16(__VA_ARGS__) -#define svuzp2_b8(...) __builtin_sve_svuzp2_b8(__VA_ARGS__) -#define svuzp2_b32(...) __builtin_sve_svuzp2_b32(__VA_ARGS__) -#define svuzp2_b64(...) __builtin_sve_svuzp2_b64(__VA_ARGS__) -#define svuzp2_b16(...) __builtin_sve_svuzp2_b16(__VA_ARGS__) -#define svwhilele_b8_s32(...) __builtin_sve_svwhilele_b8_s32(__VA_ARGS__) -#define svwhilele_b32_s32(...) __builtin_sve_svwhilele_b32_s32(__VA_ARGS__) -#define svwhilele_b64_s32(...) __builtin_sve_svwhilele_b64_s32(__VA_ARGS__) -#define svwhilele_b16_s32(...) __builtin_sve_svwhilele_b16_s32(__VA_ARGS__) -#define svwhilele_b8_s64(...) __builtin_sve_svwhilele_b8_s64(__VA_ARGS__) -#define svwhilele_b32_s64(...) __builtin_sve_svwhilele_b32_s64(__VA_ARGS__) -#define svwhilele_b64_s64(...) __builtin_sve_svwhilele_b64_s64(__VA_ARGS__) -#define svwhilele_b16_s64(...) __builtin_sve_svwhilele_b16_s64(__VA_ARGS__) -#define svwhilele_b8_u32(...) __builtin_sve_svwhilele_b8_u32(__VA_ARGS__) -#define svwhilele_b32_u32(...) __builtin_sve_svwhilele_b32_u32(__VA_ARGS__) -#define svwhilele_b64_u32(...) __builtin_sve_svwhilele_b64_u32(__VA_ARGS__) -#define svwhilele_b16_u32(...) __builtin_sve_svwhilele_b16_u32(__VA_ARGS__) -#define svwhilele_b8_u64(...) __builtin_sve_svwhilele_b8_u64(__VA_ARGS__) -#define svwhilele_b32_u64(...) __builtin_sve_svwhilele_b32_u64(__VA_ARGS__) -#define svwhilele_b64_u64(...) __builtin_sve_svwhilele_b64_u64(__VA_ARGS__) -#define svwhilele_b16_u64(...) __builtin_sve_svwhilele_b16_u64(__VA_ARGS__) -#define svwhilelt_b8_u32(...) __builtin_sve_svwhilelt_b8_u32(__VA_ARGS__) -#define svwhilelt_b32_u32(...) __builtin_sve_svwhilelt_b32_u32(__VA_ARGS__) -#define svwhilelt_b64_u32(...) __builtin_sve_svwhilelt_b64_u32(__VA_ARGS__) -#define svwhilelt_b16_u32(...) __builtin_sve_svwhilelt_b16_u32(__VA_ARGS__) -#define svwhilelt_b8_u64(...) __builtin_sve_svwhilelt_b8_u64(__VA_ARGS__) -#define svwhilelt_b32_u64(...) __builtin_sve_svwhilelt_b32_u64(__VA_ARGS__) -#define svwhilelt_b64_u64(...) __builtin_sve_svwhilelt_b64_u64(__VA_ARGS__) -#define svwhilelt_b16_u64(...) __builtin_sve_svwhilelt_b16_u64(__VA_ARGS__) -#define svwhilelt_b8_s32(...) __builtin_sve_svwhilelt_b8_s32(__VA_ARGS__) -#define svwhilelt_b32_s32(...) __builtin_sve_svwhilelt_b32_s32(__VA_ARGS__) -#define svwhilelt_b64_s32(...) __builtin_sve_svwhilelt_b64_s32(__VA_ARGS__) -#define svwhilelt_b16_s32(...) __builtin_sve_svwhilelt_b16_s32(__VA_ARGS__) -#define svwhilelt_b8_s64(...) __builtin_sve_svwhilelt_b8_s64(__VA_ARGS__) -#define svwhilelt_b32_s64(...) __builtin_sve_svwhilelt_b32_s64(__VA_ARGS__) -#define svwhilelt_b64_s64(...) __builtin_sve_svwhilelt_b64_s64(__VA_ARGS__) -#define svwhilelt_b16_s64(...) __builtin_sve_svwhilelt_b16_s64(__VA_ARGS__) -#define svwrffr(...) __builtin_sve_svwrffr(__VA_ARGS__) -#define svzip1_u8(...) __builtin_sve_svzip1_u8(__VA_ARGS__) -#define svzip1_u32(...) __builtin_sve_svzip1_u32(__VA_ARGS__) -#define svzip1_u64(...) __builtin_sve_svzip1_u64(__VA_ARGS__) -#define svzip1_u16(...) __builtin_sve_svzip1_u16(__VA_ARGS__) -#define svzip1_s8(...) __builtin_sve_svzip1_s8(__VA_ARGS__) -#define svzip1_f64(...) __builtin_sve_svzip1_f64(__VA_ARGS__) -#define svzip1_f32(...) __builtin_sve_svzip1_f32(__VA_ARGS__) -#define svzip1_f16(...) __builtin_sve_svzip1_f16(__VA_ARGS__) -#define svzip1_s32(...) __builtin_sve_svzip1_s32(__VA_ARGS__) -#define svzip1_s64(...) __builtin_sve_svzip1_s64(__VA_ARGS__) -#define svzip1_s16(...) __builtin_sve_svzip1_s16(__VA_ARGS__) -#define svzip1_b8(...) __builtin_sve_svzip1_b8(__VA_ARGS__) -#define svzip1_b32(...) __builtin_sve_svzip1_b32(__VA_ARGS__) -#define svzip1_b64(...) __builtin_sve_svzip1_b64(__VA_ARGS__) -#define svzip1_b16(...) __builtin_sve_svzip1_b16(__VA_ARGS__) -#define svzip2_u8(...) __builtin_sve_svzip2_u8(__VA_ARGS__) -#define svzip2_u32(...) __builtin_sve_svzip2_u32(__VA_ARGS__) -#define svzip2_u64(...) __builtin_sve_svzip2_u64(__VA_ARGS__) -#define svzip2_u16(...) __builtin_sve_svzip2_u16(__VA_ARGS__) -#define svzip2_s8(...) __builtin_sve_svzip2_s8(__VA_ARGS__) -#define svzip2_f64(...) __builtin_sve_svzip2_f64(__VA_ARGS__) -#define svzip2_f32(...) __builtin_sve_svzip2_f32(__VA_ARGS__) -#define svzip2_f16(...) __builtin_sve_svzip2_f16(__VA_ARGS__) -#define svzip2_s32(...) __builtin_sve_svzip2_s32(__VA_ARGS__) -#define svzip2_s64(...) __builtin_sve_svzip2_s64(__VA_ARGS__) -#define svzip2_s16(...) __builtin_sve_svzip2_s16(__VA_ARGS__) -#define svzip2_b8(...) __builtin_sve_svzip2_b8(__VA_ARGS__) -#define svzip2_b32(...) __builtin_sve_svzip2_b32(__VA_ARGS__) -#define svzip2_b64(...) __builtin_sve_svzip2_b64(__VA_ARGS__) -#define svzip2_b16(...) __builtin_sve_svzip2_b16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f64_m))) +svfloat64_t svabd_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f32_m))) +svfloat32_t svabd_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f16_m))) +svfloat16_t svabd_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f64_x))) +svfloat64_t svabd_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f32_x))) +svfloat32_t svabd_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f16_x))) +svfloat16_t svabd_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f64_z))) +svfloat64_t svabd_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f32_z))) +svfloat32_t svabd_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f16_z))) +svfloat16_t svabd_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s8_m))) +svint8_t svabd_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s32_m))) +svint32_t svabd_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s64_m))) +svint64_t svabd_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s16_m))) +svint16_t svabd_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s8_x))) +svint8_t svabd_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s32_x))) +svint32_t svabd_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s64_x))) +svint64_t svabd_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s16_x))) +svint16_t svabd_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s8_z))) +svint8_t svabd_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s32_z))) +svint32_t svabd_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s64_z))) +svint64_t svabd_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s16_z))) +svint16_t svabd_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u8_m))) +svuint8_t svabd_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u32_m))) +svuint32_t svabd_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u64_m))) +svuint64_t svabd_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u16_m))) +svuint16_t svabd_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u8_x))) +svuint8_t svabd_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u32_x))) +svuint32_t svabd_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u64_x))) +svuint64_t svabd_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u16_x))) +svuint16_t svabd_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u8_z))) +svuint8_t svabd_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u32_z))) +svuint32_t svabd_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u64_z))) +svuint64_t svabd_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u16_z))) +svuint16_t svabd_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f64_m))) +svfloat64_t svabd_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f32_m))) +svfloat32_t svabd_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f16_m))) +svfloat16_t svabd_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f64_x))) +svfloat64_t svabd_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f32_x))) +svfloat32_t svabd_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f16_x))) +svfloat16_t svabd_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f64_z))) +svfloat64_t svabd_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f32_z))) +svfloat32_t svabd_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f16_z))) +svfloat16_t svabd_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s8_m))) +svint8_t svabd_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s32_m))) +svint32_t svabd_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s64_m))) +svint64_t svabd_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s16_m))) +svint16_t svabd_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s8_x))) +svint8_t svabd_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s32_x))) +svint32_t svabd_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s64_x))) +svint64_t svabd_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s16_x))) +svint16_t svabd_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s8_z))) +svint8_t svabd_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s32_z))) +svint32_t svabd_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s64_z))) +svint64_t svabd_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s16_z))) +svint16_t svabd_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u8_m))) +svuint8_t svabd_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u32_m))) +svuint32_t svabd_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u64_m))) +svuint64_t svabd_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u16_m))) +svuint16_t svabd_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u8_x))) +svuint8_t svabd_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u32_x))) +svuint32_t svabd_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u64_x))) +svuint64_t svabd_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u16_x))) +svuint16_t svabd_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u8_z))) +svuint8_t svabd_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u32_z))) +svuint32_t svabd_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u64_z))) +svuint64_t svabd_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u16_z))) +svuint16_t svabd_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f64_m))) +svfloat64_t svabs_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f32_m))) +svfloat32_t svabs_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f16_m))) +svfloat16_t svabs_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f64_x))) +svfloat64_t svabs_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f32_x))) +svfloat32_t svabs_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f16_x))) +svfloat16_t svabs_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f64_z))) +svfloat64_t svabs_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f32_z))) +svfloat32_t svabs_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f16_z))) +svfloat16_t svabs_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s8_m))) +svint8_t svabs_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s32_m))) +svint32_t svabs_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s64_m))) +svint64_t svabs_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s16_m))) +svint16_t svabs_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s8_x))) +svint8_t svabs_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s32_x))) +svint32_t svabs_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s64_x))) +svint64_t svabs_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s16_x))) +svint16_t svabs_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s8_z))) +svint8_t svabs_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s32_z))) +svint32_t svabs_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s64_z))) +svint64_t svabs_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s16_z))) +svint16_t svabs_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacge_n_f64))) +svbool_t svacge_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacge_n_f32))) +svbool_t svacge_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacge_n_f16))) +svbool_t svacge_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacge_f64))) +svbool_t svacge_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacge_f32))) +svbool_t svacge_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacge_f16))) +svbool_t svacge_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacgt_n_f64))) +svbool_t svacgt_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacgt_n_f32))) +svbool_t svacgt_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacgt_n_f16))) +svbool_t svacgt_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacgt_f64))) +svbool_t svacgt_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacgt_f32))) +svbool_t svacgt_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacgt_f16))) +svbool_t svacgt_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacle_n_f64))) +svbool_t svacle_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacle_n_f32))) +svbool_t svacle_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacle_n_f16))) +svbool_t svacle_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacle_f64))) +svbool_t svacle_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacle_f32))) +svbool_t svacle_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacle_f16))) +svbool_t svacle_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaclt_n_f64))) +svbool_t svaclt_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaclt_n_f32))) +svbool_t svaclt_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaclt_n_f16))) +svbool_t svaclt_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaclt_f64))) +svbool_t svaclt_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaclt_f32))) +svbool_t svaclt_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaclt_f16))) +svbool_t svaclt_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f64_m))) +svfloat64_t svadd_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f32_m))) +svfloat32_t svadd_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f16_m))) +svfloat16_t svadd_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f64_x))) +svfloat64_t svadd_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f32_x))) +svfloat32_t svadd_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f16_x))) +svfloat16_t svadd_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f64_z))) +svfloat64_t svadd_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f32_z))) +svfloat32_t svadd_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f16_z))) +svfloat16_t svadd_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u8_m))) +svuint8_t svadd_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u32_m))) +svuint32_t svadd_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u64_m))) +svuint64_t svadd_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u16_m))) +svuint16_t svadd_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s8_m))) +svint8_t svadd_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s32_m))) +svint32_t svadd_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s64_m))) +svint64_t svadd_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s16_m))) +svint16_t svadd_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u8_x))) +svuint8_t svadd_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u32_x))) +svuint32_t svadd_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u64_x))) +svuint64_t svadd_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u16_x))) +svuint16_t svadd_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s8_x))) +svint8_t svadd_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s32_x))) +svint32_t svadd_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s64_x))) +svint64_t svadd_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s16_x))) +svint16_t svadd_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u8_z))) +svuint8_t svadd_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u32_z))) +svuint32_t svadd_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u64_z))) +svuint64_t svadd_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u16_z))) +svuint16_t svadd_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s8_z))) +svint8_t svadd_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s32_z))) +svint32_t svadd_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s64_z))) +svint64_t svadd_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s16_z))) +svint16_t svadd_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f64_m))) +svfloat64_t svadd_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f32_m))) +svfloat32_t svadd_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f16_m))) +svfloat16_t svadd_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f64_x))) +svfloat64_t svadd_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f32_x))) +svfloat32_t svadd_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f16_x))) +svfloat16_t svadd_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f64_z))) +svfloat64_t svadd_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f32_z))) +svfloat32_t svadd_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f16_z))) +svfloat16_t svadd_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u8_m))) +svuint8_t svadd_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u32_m))) +svuint32_t svadd_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u64_m))) +svuint64_t svadd_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u16_m))) +svuint16_t svadd_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s8_m))) +svint8_t svadd_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s32_m))) +svint32_t svadd_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s64_m))) +svint64_t svadd_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s16_m))) +svint16_t svadd_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u8_x))) +svuint8_t svadd_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u32_x))) +svuint32_t svadd_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u64_x))) +svuint64_t svadd_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u16_x))) +svuint16_t svadd_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s8_x))) +svint8_t svadd_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s32_x))) +svint32_t svadd_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s64_x))) +svint64_t svadd_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s16_x))) +svint16_t svadd_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u8_z))) +svuint8_t svadd_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u32_z))) +svuint32_t svadd_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u64_z))) +svuint64_t svadd_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u16_z))) +svuint16_t svadd_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s8_z))) +svint8_t svadd_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s32_z))) +svint32_t svadd_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s64_z))) +svint64_t svadd_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s16_z))) +svint16_t svadd_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadda_f64))) +float64_t svadda_f64(svbool_t, float64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadda_f32))) +float32_t svadda_f32(svbool_t, float32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadda_f16))) +float16_t svadda_f16(svbool_t, float16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_s8))) +int64_t svaddv_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_s32))) +int64_t svaddv_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_s64))) +int64_t svaddv_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_s16))) +int64_t svaddv_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_u8))) +uint64_t svaddv_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_u32))) +uint64_t svaddv_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_u64))) +uint64_t svaddv_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_u16))) +uint64_t svaddv_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_f64))) +float64_t svaddv_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_f32))) +float32_t svaddv_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_f16))) +float16_t svaddv_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrb_u32base_u32offset))) +svuint32_t svadrb_u32base_u32offset(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrb_u64base_u64offset))) +svuint64_t svadrb_u64base_u64offset(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrb_u32base_s32offset))) +svuint32_t svadrb_u32base_s32offset(svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrb_u64base_s64offset))) +svuint64_t svadrb_u64base_s64offset(svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrd_u32base_u32index))) +svuint32_t svadrd_u32base_u32index(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrd_u64base_u64index))) +svuint64_t svadrd_u64base_u64index(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrd_u32base_s32index))) +svuint32_t svadrd_u32base_s32index(svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrd_u64base_s64index))) +svuint64_t svadrd_u64base_s64index(svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrh_u32base_u32index))) +svuint32_t svadrh_u32base_u32index(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrh_u64base_u64index))) +svuint64_t svadrh_u64base_u64index(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrh_u32base_s32index))) +svuint32_t svadrh_u32base_s32index(svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrh_u64base_s64index))) +svuint64_t svadrh_u64base_s64index(svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrw_u32base_u32index))) +svuint32_t svadrw_u32base_u32index(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrw_u64base_u64index))) +svuint64_t svadrw_u64base_u64index(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrw_u32base_s32index))) +svuint32_t svadrw_u32base_s32index(svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrw_u64base_s64index))) +svuint64_t svadrw_u64base_s64index(svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_b_z))) +svbool_t svand_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u8_m))) +svuint8_t svand_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u32_m))) +svuint32_t svand_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u64_m))) +svuint64_t svand_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u16_m))) +svuint16_t svand_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s8_m))) +svint8_t svand_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s32_m))) +svint32_t svand_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s64_m))) +svint64_t svand_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s16_m))) +svint16_t svand_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u8_x))) +svuint8_t svand_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u32_x))) +svuint32_t svand_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u64_x))) +svuint64_t svand_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u16_x))) +svuint16_t svand_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s8_x))) +svint8_t svand_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s32_x))) +svint32_t svand_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s64_x))) +svint64_t svand_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s16_x))) +svint16_t svand_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u8_z))) +svuint8_t svand_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u32_z))) +svuint32_t svand_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u64_z))) +svuint64_t svand_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u16_z))) +svuint16_t svand_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s8_z))) +svint8_t svand_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s32_z))) +svint32_t svand_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s64_z))) +svint64_t svand_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s16_z))) +svint16_t svand_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u8_m))) +svuint8_t svand_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u32_m))) +svuint32_t svand_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u64_m))) +svuint64_t svand_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u16_m))) +svuint16_t svand_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s8_m))) +svint8_t svand_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s32_m))) +svint32_t svand_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s64_m))) +svint64_t svand_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s16_m))) +svint16_t svand_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u8_x))) +svuint8_t svand_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u32_x))) +svuint32_t svand_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u64_x))) +svuint64_t svand_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u16_x))) +svuint16_t svand_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s8_x))) +svint8_t svand_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s32_x))) +svint32_t svand_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s64_x))) +svint64_t svand_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s16_x))) +svint16_t svand_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u8_z))) +svuint8_t svand_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u32_z))) +svuint32_t svand_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u64_z))) +svuint64_t svand_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u16_z))) +svuint16_t svand_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s8_z))) +svint8_t svand_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s32_z))) +svint32_t svand_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s64_z))) +svint64_t svand_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s16_z))) +svint16_t svand_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_u8))) +uint8_t svandv_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_u32))) +uint32_t svandv_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_u64))) +uint64_t svandv_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_u16))) +uint16_t svandv_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_s8))) +int8_t svandv_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_s32))) +int32_t svandv_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_s64))) +int64_t svandv_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_s16))) +int16_t svandv_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s8_m))) +svint8_t svasr_n_s8_m(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s32_m))) +svint32_t svasr_n_s32_m(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s64_m))) +svint64_t svasr_n_s64_m(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s16_m))) +svint16_t svasr_n_s16_m(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s8_x))) +svint8_t svasr_n_s8_x(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s32_x))) +svint32_t svasr_n_s32_x(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s64_x))) +svint64_t svasr_n_s64_x(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s16_x))) +svint16_t svasr_n_s16_x(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s8_z))) +svint8_t svasr_n_s8_z(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s32_z))) +svint32_t svasr_n_s32_z(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s64_z))) +svint64_t svasr_n_s64_z(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s16_z))) +svint16_t svasr_n_s16_z(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s8_m))) +svint8_t svasr_s8_m(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s32_m))) +svint32_t svasr_s32_m(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s64_m))) +svint64_t svasr_s64_m(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s16_m))) +svint16_t svasr_s16_m(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s8_x))) +svint8_t svasr_s8_x(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s32_x))) +svint32_t svasr_s32_x(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s64_x))) +svint64_t svasr_s64_x(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s16_x))) +svint16_t svasr_s16_x(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s8_z))) +svint8_t svasr_s8_z(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s32_z))) +svint32_t svasr_s32_z(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s64_z))) +svint64_t svasr_s64_z(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s16_z))) +svint16_t svasr_s16_z(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s8_m))) +svint8_t svasr_wide_n_s8_m(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s32_m))) +svint32_t svasr_wide_n_s32_m(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s16_m))) +svint16_t svasr_wide_n_s16_m(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s8_x))) +svint8_t svasr_wide_n_s8_x(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s32_x))) +svint32_t svasr_wide_n_s32_x(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s16_x))) +svint16_t svasr_wide_n_s16_x(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s8_z))) +svint8_t svasr_wide_n_s8_z(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s32_z))) +svint32_t svasr_wide_n_s32_z(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s16_z))) +svint16_t svasr_wide_n_s16_z(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s8_m))) +svint8_t svasr_wide_s8_m(svbool_t, svint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s32_m))) +svint32_t svasr_wide_s32_m(svbool_t, svint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s16_m))) +svint16_t svasr_wide_s16_m(svbool_t, svint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s8_x))) +svint8_t svasr_wide_s8_x(svbool_t, svint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s32_x))) +svint32_t svasr_wide_s32_x(svbool_t, svint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s16_x))) +svint16_t svasr_wide_s16_x(svbool_t, svint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s8_z))) +svint8_t svasr_wide_s8_z(svbool_t, svint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s32_z))) +svint32_t svasr_wide_s32_z(svbool_t, svint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s16_z))) +svint16_t svasr_wide_s16_z(svbool_t, svint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s8_m))) +svint8_t svasrd_n_s8_m(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s32_m))) +svint32_t svasrd_n_s32_m(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s64_m))) +svint64_t svasrd_n_s64_m(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s16_m))) +svint16_t svasrd_n_s16_m(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s8_x))) +svint8_t svasrd_n_s8_x(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s32_x))) +svint32_t svasrd_n_s32_x(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s64_x))) +svint64_t svasrd_n_s64_x(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s16_x))) +svint16_t svasrd_n_s16_x(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s8_z))) +svint8_t svasrd_n_s8_z(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s32_z))) +svint32_t svasrd_n_s32_z(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s64_z))) +svint64_t svasrd_n_s64_z(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s16_z))) +svint16_t svasrd_n_s16_z(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_b_z))) +svbool_t svbic_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u8_m))) +svuint8_t svbic_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u32_m))) +svuint32_t svbic_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u64_m))) +svuint64_t svbic_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u16_m))) +svuint16_t svbic_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s8_m))) +svint8_t svbic_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s32_m))) +svint32_t svbic_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s64_m))) +svint64_t svbic_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s16_m))) +svint16_t svbic_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u8_x))) +svuint8_t svbic_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u32_x))) +svuint32_t svbic_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u64_x))) +svuint64_t svbic_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u16_x))) +svuint16_t svbic_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s8_x))) +svint8_t svbic_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s32_x))) +svint32_t svbic_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s64_x))) +svint64_t svbic_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s16_x))) +svint16_t svbic_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u8_z))) +svuint8_t svbic_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u32_z))) +svuint32_t svbic_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u64_z))) +svuint64_t svbic_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u16_z))) +svuint16_t svbic_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s8_z))) +svint8_t svbic_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s32_z))) +svint32_t svbic_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s64_z))) +svint64_t svbic_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s16_z))) +svint16_t svbic_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u8_m))) +svuint8_t svbic_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u32_m))) +svuint32_t svbic_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u64_m))) +svuint64_t svbic_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u16_m))) +svuint16_t svbic_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s8_m))) +svint8_t svbic_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s32_m))) +svint32_t svbic_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s64_m))) +svint64_t svbic_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s16_m))) +svint16_t svbic_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u8_x))) +svuint8_t svbic_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u32_x))) +svuint32_t svbic_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u64_x))) +svuint64_t svbic_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u16_x))) +svuint16_t svbic_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s8_x))) +svint8_t svbic_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s32_x))) +svint32_t svbic_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s64_x))) +svint64_t svbic_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s16_x))) +svint16_t svbic_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u8_z))) +svuint8_t svbic_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u32_z))) +svuint32_t svbic_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u64_z))) +svuint64_t svbic_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u16_z))) +svuint16_t svbic_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s8_z))) +svint8_t svbic_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s32_z))) +svint32_t svbic_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s64_z))) +svint64_t svbic_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s16_z))) +svint16_t svbic_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrka_b_m))) +svbool_t svbrka_b_m(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrka_b_z))) +svbool_t svbrka_b_z(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrkb_b_m))) +svbool_t svbrkb_b_m(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrkb_b_z))) +svbool_t svbrkb_b_z(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrkn_b_z))) +svbool_t svbrkn_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrkpa_b_z))) +svbool_t svbrkpa_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrkpb_b_z))) +svbool_t svbrkpb_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f64_m))) +svfloat64_t svcadd_f64_m(svbool_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f32_m))) +svfloat32_t svcadd_f32_m(svbool_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f16_m))) +svfloat16_t svcadd_f16_m(svbool_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f64_x))) +svfloat64_t svcadd_f64_x(svbool_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f32_x))) +svfloat32_t svcadd_f32_x(svbool_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f16_x))) +svfloat16_t svcadd_f16_x(svbool_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f64_z))) +svfloat64_t svcadd_f64_z(svbool_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f32_z))) +svfloat32_t svcadd_f32_z(svbool_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f16_z))) +svfloat16_t svcadd_f16_z(svbool_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_u8))) +uint8_t svclasta_n_u8(svbool_t, uint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_u32))) +uint32_t svclasta_n_u32(svbool_t, uint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_u64))) +uint64_t svclasta_n_u64(svbool_t, uint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_u16))) +uint16_t svclasta_n_u16(svbool_t, uint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_s8))) +int8_t svclasta_n_s8(svbool_t, int8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_f64))) +float64_t svclasta_n_f64(svbool_t, float64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_f32))) +float32_t svclasta_n_f32(svbool_t, float32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_f16))) +float16_t svclasta_n_f16(svbool_t, float16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_s32))) +int32_t svclasta_n_s32(svbool_t, int32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_s64))) +int64_t svclasta_n_s64(svbool_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_s16))) +int16_t svclasta_n_s16(svbool_t, int16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_u8))) +svuint8_t svclasta_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_u32))) +svuint32_t svclasta_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_u64))) +svuint64_t svclasta_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_u16))) +svuint16_t svclasta_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_s8))) +svint8_t svclasta_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_f64))) +svfloat64_t svclasta_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_f32))) +svfloat32_t svclasta_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_f16))) +svfloat16_t svclasta_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_s32))) +svint32_t svclasta_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_s64))) +svint64_t svclasta_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_s16))) +svint16_t svclasta_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_u8))) +uint8_t svclastb_n_u8(svbool_t, uint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_u32))) +uint32_t svclastb_n_u32(svbool_t, uint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_u64))) +uint64_t svclastb_n_u64(svbool_t, uint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_u16))) +uint16_t svclastb_n_u16(svbool_t, uint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_s8))) +int8_t svclastb_n_s8(svbool_t, int8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_f64))) +float64_t svclastb_n_f64(svbool_t, float64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_f32))) +float32_t svclastb_n_f32(svbool_t, float32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_f16))) +float16_t svclastb_n_f16(svbool_t, float16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_s32))) +int32_t svclastb_n_s32(svbool_t, int32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_s64))) +int64_t svclastb_n_s64(svbool_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_s16))) +int16_t svclastb_n_s16(svbool_t, int16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_u8))) +svuint8_t svclastb_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_u32))) +svuint32_t svclastb_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_u64))) +svuint64_t svclastb_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_u16))) +svuint16_t svclastb_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_s8))) +svint8_t svclastb_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_f64))) +svfloat64_t svclastb_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_f32))) +svfloat32_t svclastb_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_f16))) +svfloat16_t svclastb_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_s32))) +svint32_t svclastb_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_s64))) +svint64_t svclastb_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_s16))) +svint16_t svclastb_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s8_m))) +svuint8_t svcls_s8_m(svuint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s32_m))) +svuint32_t svcls_s32_m(svuint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s64_m))) +svuint64_t svcls_s64_m(svuint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s16_m))) +svuint16_t svcls_s16_m(svuint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s8_x))) +svuint8_t svcls_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s32_x))) +svuint32_t svcls_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s64_x))) +svuint64_t svcls_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s16_x))) +svuint16_t svcls_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s8_z))) +svuint8_t svcls_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s32_z))) +svuint32_t svcls_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s64_z))) +svuint64_t svcls_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s16_z))) +svuint16_t svcls_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u8_m))) +svuint8_t svclz_u8_m(svuint8_t, svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u32_m))) +svuint32_t svclz_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u64_m))) +svuint64_t svclz_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u16_m))) +svuint16_t svclz_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s8_m))) +svuint8_t svclz_s8_m(svuint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s32_m))) +svuint32_t svclz_s32_m(svuint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s64_m))) +svuint64_t svclz_s64_m(svuint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s16_m))) +svuint16_t svclz_s16_m(svuint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u8_x))) +svuint8_t svclz_u8_x(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u32_x))) +svuint32_t svclz_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u64_x))) +svuint64_t svclz_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u16_x))) +svuint16_t svclz_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s8_x))) +svuint8_t svclz_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s32_x))) +svuint32_t svclz_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s64_x))) +svuint64_t svclz_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s16_x))) +svuint16_t svclz_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u8_z))) +svuint8_t svclz_u8_z(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u32_z))) +svuint32_t svclz_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u64_z))) +svuint64_t svclz_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u16_z))) +svuint16_t svclz_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s8_z))) +svuint8_t svclz_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s32_z))) +svuint32_t svclz_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s64_z))) +svuint64_t svclz_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s16_z))) +svuint16_t svclz_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f64_m))) +svfloat64_t svcmla_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f32_m))) +svfloat32_t svcmla_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f16_m))) +svfloat16_t svcmla_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f64_x))) +svfloat64_t svcmla_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f32_x))) +svfloat32_t svcmla_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f16_x))) +svfloat16_t svcmla_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f64_z))) +svfloat64_t svcmla_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f32_z))) +svfloat32_t svcmla_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f16_z))) +svfloat16_t svcmla_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_lane_f32))) +svfloat32_t svcmla_lane_f32(svfloat32_t, svfloat32_t, svfloat32_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_lane_f16))) +svfloat16_t svcmla_lane_f16(svfloat16_t, svfloat16_t, svfloat16_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_f64))) +svbool_t svcmpeq_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_f32))) +svbool_t svcmpeq_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_f16))) +svbool_t svcmpeq_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_u8))) +svbool_t svcmpeq_n_u8(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_u32))) +svbool_t svcmpeq_n_u32(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_u64))) +svbool_t svcmpeq_n_u64(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_u16))) +svbool_t svcmpeq_n_u16(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_s8))) +svbool_t svcmpeq_n_s8(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_s32))) +svbool_t svcmpeq_n_s32(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_s64))) +svbool_t svcmpeq_n_s64(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_s16))) +svbool_t svcmpeq_n_s16(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_u8))) +svbool_t svcmpeq_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_u32))) +svbool_t svcmpeq_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_u64))) +svbool_t svcmpeq_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_u16))) +svbool_t svcmpeq_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_s8))) +svbool_t svcmpeq_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_s32))) +svbool_t svcmpeq_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_s64))) +svbool_t svcmpeq_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_s16))) +svbool_t svcmpeq_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_f64))) +svbool_t svcmpeq_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_f32))) +svbool_t svcmpeq_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_f16))) +svbool_t svcmpeq_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_wide_n_s8))) +svbool_t svcmpeq_wide_n_s8(svbool_t, svint8_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_wide_n_s32))) +svbool_t svcmpeq_wide_n_s32(svbool_t, svint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_wide_n_s16))) +svbool_t svcmpeq_wide_n_s16(svbool_t, svint16_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_wide_s8))) +svbool_t svcmpeq_wide_s8(svbool_t, svint8_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_wide_s32))) +svbool_t svcmpeq_wide_s32(svbool_t, svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_wide_s16))) +svbool_t svcmpeq_wide_s16(svbool_t, svint16_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_f64))) +svbool_t svcmpge_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_f32))) +svbool_t svcmpge_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_f16))) +svbool_t svcmpge_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_s8))) +svbool_t svcmpge_n_s8(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_s32))) +svbool_t svcmpge_n_s32(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_s64))) +svbool_t svcmpge_n_s64(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_s16))) +svbool_t svcmpge_n_s16(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_u8))) +svbool_t svcmpge_n_u8(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_u32))) +svbool_t svcmpge_n_u32(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_u64))) +svbool_t svcmpge_n_u64(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_u16))) +svbool_t svcmpge_n_u16(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_s8))) +svbool_t svcmpge_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_s32))) +svbool_t svcmpge_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_s64))) +svbool_t svcmpge_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_s16))) +svbool_t svcmpge_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_f64))) +svbool_t svcmpge_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_f32))) +svbool_t svcmpge_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_f16))) +svbool_t svcmpge_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_u8))) +svbool_t svcmpge_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_u32))) +svbool_t svcmpge_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_u64))) +svbool_t svcmpge_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_u16))) +svbool_t svcmpge_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_n_s8))) +svbool_t svcmpge_wide_n_s8(svbool_t, svint8_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_n_s32))) +svbool_t svcmpge_wide_n_s32(svbool_t, svint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_n_s16))) +svbool_t svcmpge_wide_n_s16(svbool_t, svint16_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_n_u8))) +svbool_t svcmpge_wide_n_u8(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_n_u32))) +svbool_t svcmpge_wide_n_u32(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_n_u16))) +svbool_t svcmpge_wide_n_u16(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_s8))) +svbool_t svcmpge_wide_s8(svbool_t, svint8_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_s32))) +svbool_t svcmpge_wide_s32(svbool_t, svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_s16))) +svbool_t svcmpge_wide_s16(svbool_t, svint16_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_u8))) +svbool_t svcmpge_wide_u8(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_u32))) +svbool_t svcmpge_wide_u32(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_u16))) +svbool_t svcmpge_wide_u16(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_f64))) +svbool_t svcmpgt_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_f32))) +svbool_t svcmpgt_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_f16))) +svbool_t svcmpgt_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_s8))) +svbool_t svcmpgt_n_s8(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_s32))) +svbool_t svcmpgt_n_s32(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_s64))) +svbool_t svcmpgt_n_s64(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_s16))) +svbool_t svcmpgt_n_s16(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_u8))) +svbool_t svcmpgt_n_u8(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_u32))) +svbool_t svcmpgt_n_u32(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_u64))) +svbool_t svcmpgt_n_u64(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_u16))) +svbool_t svcmpgt_n_u16(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_s8))) +svbool_t svcmpgt_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_s32))) +svbool_t svcmpgt_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_s64))) +svbool_t svcmpgt_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_s16))) +svbool_t svcmpgt_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_f64))) +svbool_t svcmpgt_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_f32))) +svbool_t svcmpgt_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_f16))) +svbool_t svcmpgt_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_u8))) +svbool_t svcmpgt_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_u32))) +svbool_t svcmpgt_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_u64))) +svbool_t svcmpgt_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_u16))) +svbool_t svcmpgt_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_n_s8))) +svbool_t svcmpgt_wide_n_s8(svbool_t, svint8_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_n_s32))) +svbool_t svcmpgt_wide_n_s32(svbool_t, svint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_n_s16))) +svbool_t svcmpgt_wide_n_s16(svbool_t, svint16_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_n_u8))) +svbool_t svcmpgt_wide_n_u8(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_n_u32))) +svbool_t svcmpgt_wide_n_u32(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_n_u16))) +svbool_t svcmpgt_wide_n_u16(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_s8))) +svbool_t svcmpgt_wide_s8(svbool_t, svint8_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_s32))) +svbool_t svcmpgt_wide_s32(svbool_t, svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_s16))) +svbool_t svcmpgt_wide_s16(svbool_t, svint16_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_u8))) +svbool_t svcmpgt_wide_u8(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_u32))) +svbool_t svcmpgt_wide_u32(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_u16))) +svbool_t svcmpgt_wide_u16(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_f64))) +svbool_t svcmple_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_f32))) +svbool_t svcmple_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_f16))) +svbool_t svcmple_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_s8))) +svbool_t svcmple_n_s8(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_s32))) +svbool_t svcmple_n_s32(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_s64))) +svbool_t svcmple_n_s64(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_s16))) +svbool_t svcmple_n_s16(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_u8))) +svbool_t svcmple_n_u8(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_u32))) +svbool_t svcmple_n_u32(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_u64))) +svbool_t svcmple_n_u64(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_u16))) +svbool_t svcmple_n_u16(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_s8))) +svbool_t svcmple_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_s32))) +svbool_t svcmple_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_s64))) +svbool_t svcmple_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_s16))) +svbool_t svcmple_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_f64))) +svbool_t svcmple_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_f32))) +svbool_t svcmple_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_f16))) +svbool_t svcmple_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_u8))) +svbool_t svcmple_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_u32))) +svbool_t svcmple_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_u64))) +svbool_t svcmple_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_u16))) +svbool_t svcmple_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_n_s8))) +svbool_t svcmple_wide_n_s8(svbool_t, svint8_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_n_s32))) +svbool_t svcmple_wide_n_s32(svbool_t, svint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_n_s16))) +svbool_t svcmple_wide_n_s16(svbool_t, svint16_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_n_u8))) +svbool_t svcmple_wide_n_u8(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_n_u32))) +svbool_t svcmple_wide_n_u32(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_n_u16))) +svbool_t svcmple_wide_n_u16(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_s8))) +svbool_t svcmple_wide_s8(svbool_t, svint8_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_s32))) +svbool_t svcmple_wide_s32(svbool_t, svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_s16))) +svbool_t svcmple_wide_s16(svbool_t, svint16_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_u8))) +svbool_t svcmple_wide_u8(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_u32))) +svbool_t svcmple_wide_u32(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_u16))) +svbool_t svcmple_wide_u16(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_u8))) +svbool_t svcmplt_n_u8(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_u32))) +svbool_t svcmplt_n_u32(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_u64))) +svbool_t svcmplt_n_u64(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_u16))) +svbool_t svcmplt_n_u16(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_f64))) +svbool_t svcmplt_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_f32))) +svbool_t svcmplt_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_f16))) +svbool_t svcmplt_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_s8))) +svbool_t svcmplt_n_s8(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_s32))) +svbool_t svcmplt_n_s32(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_s64))) +svbool_t svcmplt_n_s64(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_s16))) +svbool_t svcmplt_n_s16(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_u8))) +svbool_t svcmplt_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_u32))) +svbool_t svcmplt_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_u64))) +svbool_t svcmplt_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_u16))) +svbool_t svcmplt_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_s8))) +svbool_t svcmplt_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_s32))) +svbool_t svcmplt_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_s64))) +svbool_t svcmplt_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_s16))) +svbool_t svcmplt_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_f64))) +svbool_t svcmplt_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_f32))) +svbool_t svcmplt_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_f16))) +svbool_t svcmplt_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_n_u8))) +svbool_t svcmplt_wide_n_u8(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_n_u32))) +svbool_t svcmplt_wide_n_u32(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_n_u16))) +svbool_t svcmplt_wide_n_u16(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_n_s8))) +svbool_t svcmplt_wide_n_s8(svbool_t, svint8_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_n_s32))) +svbool_t svcmplt_wide_n_s32(svbool_t, svint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_n_s16))) +svbool_t svcmplt_wide_n_s16(svbool_t, svint16_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_u8))) +svbool_t svcmplt_wide_u8(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_u32))) +svbool_t svcmplt_wide_u32(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_u16))) +svbool_t svcmplt_wide_u16(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_s8))) +svbool_t svcmplt_wide_s8(svbool_t, svint8_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_s32))) +svbool_t svcmplt_wide_s32(svbool_t, svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_s16))) +svbool_t svcmplt_wide_s16(svbool_t, svint16_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_f64))) +svbool_t svcmpne_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_f32))) +svbool_t svcmpne_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_f16))) +svbool_t svcmpne_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_u8))) +svbool_t svcmpne_n_u8(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_u32))) +svbool_t svcmpne_n_u32(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_u64))) +svbool_t svcmpne_n_u64(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_u16))) +svbool_t svcmpne_n_u16(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_s8))) +svbool_t svcmpne_n_s8(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_s32))) +svbool_t svcmpne_n_s32(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_s64))) +svbool_t svcmpne_n_s64(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_s16))) +svbool_t svcmpne_n_s16(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_u8))) +svbool_t svcmpne_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_u32))) +svbool_t svcmpne_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_u64))) +svbool_t svcmpne_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_u16))) +svbool_t svcmpne_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_s8))) +svbool_t svcmpne_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_s32))) +svbool_t svcmpne_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_s64))) +svbool_t svcmpne_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_s16))) +svbool_t svcmpne_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_f64))) +svbool_t svcmpne_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_f32))) +svbool_t svcmpne_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_f16))) +svbool_t svcmpne_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_wide_n_s8))) +svbool_t svcmpne_wide_n_s8(svbool_t, svint8_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_wide_n_s32))) +svbool_t svcmpne_wide_n_s32(svbool_t, svint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_wide_n_s16))) +svbool_t svcmpne_wide_n_s16(svbool_t, svint16_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_wide_s8))) +svbool_t svcmpne_wide_s8(svbool_t, svint8_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_wide_s32))) +svbool_t svcmpne_wide_s32(svbool_t, svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_wide_s16))) +svbool_t svcmpne_wide_s16(svbool_t, svint16_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpuo_n_f64))) +svbool_t svcmpuo_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpuo_n_f32))) +svbool_t svcmpuo_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpuo_n_f16))) +svbool_t svcmpuo_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpuo_f64))) +svbool_t svcmpuo_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpuo_f32))) +svbool_t svcmpuo_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpuo_f16))) +svbool_t svcmpuo_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u8_m))) +svuint8_t svcnot_u8_m(svuint8_t, svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u32_m))) +svuint32_t svcnot_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u64_m))) +svuint64_t svcnot_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u16_m))) +svuint16_t svcnot_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s8_m))) +svint8_t svcnot_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s32_m))) +svint32_t svcnot_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s64_m))) +svint64_t svcnot_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s16_m))) +svint16_t svcnot_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u8_x))) +svuint8_t svcnot_u8_x(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u32_x))) +svuint32_t svcnot_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u64_x))) +svuint64_t svcnot_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u16_x))) +svuint16_t svcnot_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s8_x))) +svint8_t svcnot_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s32_x))) +svint32_t svcnot_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s64_x))) +svint64_t svcnot_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s16_x))) +svint16_t svcnot_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u8_z))) +svuint8_t svcnot_u8_z(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u32_z))) +svuint32_t svcnot_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u64_z))) +svuint64_t svcnot_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u16_z))) +svuint16_t svcnot_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s8_z))) +svint8_t svcnot_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s32_z))) +svint32_t svcnot_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s64_z))) +svint64_t svcnot_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s16_z))) +svint16_t svcnot_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u8_m))) +svuint8_t svcnt_u8_m(svuint8_t, svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u32_m))) +svuint32_t svcnt_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u64_m))) +svuint64_t svcnt_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u16_m))) +svuint16_t svcnt_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s8_m))) +svuint8_t svcnt_s8_m(svuint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f64_m))) +svuint64_t svcnt_f64_m(svuint64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f32_m))) +svuint32_t svcnt_f32_m(svuint32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f16_m))) +svuint16_t svcnt_f16_m(svuint16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s32_m))) +svuint32_t svcnt_s32_m(svuint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s64_m))) +svuint64_t svcnt_s64_m(svuint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s16_m))) +svuint16_t svcnt_s16_m(svuint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u8_x))) +svuint8_t svcnt_u8_x(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u32_x))) +svuint32_t svcnt_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u64_x))) +svuint64_t svcnt_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u16_x))) +svuint16_t svcnt_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s8_x))) +svuint8_t svcnt_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f64_x))) +svuint64_t svcnt_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f32_x))) +svuint32_t svcnt_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f16_x))) +svuint16_t svcnt_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s32_x))) +svuint32_t svcnt_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s64_x))) +svuint64_t svcnt_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s16_x))) +svuint16_t svcnt_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u8_z))) +svuint8_t svcnt_u8_z(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u32_z))) +svuint32_t svcnt_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u64_z))) +svuint64_t svcnt_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u16_z))) +svuint16_t svcnt_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s8_z))) +svuint8_t svcnt_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f64_z))) +svuint64_t svcnt_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f32_z))) +svuint32_t svcnt_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f16_z))) +svuint16_t svcnt_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s32_z))) +svuint32_t svcnt_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s64_z))) +svuint64_t svcnt_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s16_z))) +svuint16_t svcnt_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntb))) +uint64_t svcntb(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntb_pat))) +uint64_t svcntb_pat(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntd))) +uint64_t svcntd(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntd_pat))) +uint64_t svcntd_pat(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnth))) +uint64_t svcnth(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnth_pat))) +uint64_t svcnth_pat(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntp_b8))) +uint64_t svcntp_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntp_b32))) +uint64_t svcntp_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntp_b64))) +uint64_t svcntp_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntp_b16))) +uint64_t svcntp_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntw))) +uint64_t svcntw(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntw_pat))) +uint64_t svcntw_pat(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcompact_u32))) +svuint32_t svcompact_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcompact_u64))) +svuint64_t svcompact_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcompact_f64))) +svfloat64_t svcompact_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcompact_f32))) +svfloat32_t svcompact_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcompact_s32))) +svint32_t svcompact_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcompact_s64))) +svint64_t svcompact_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_u8))) +svuint8x2_t svcreate2_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_u32))) +svuint32x2_t svcreate2_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_u64))) +svuint64x2_t svcreate2_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_u16))) +svuint16x2_t svcreate2_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_s8))) +svint8x2_t svcreate2_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_f64))) +svfloat64x2_t svcreate2_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_f32))) +svfloat32x2_t svcreate2_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_f16))) +svfloat16x2_t svcreate2_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_s32))) +svint32x2_t svcreate2_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_s64))) +svint64x2_t svcreate2_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_s16))) +svint16x2_t svcreate2_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_u8))) +svuint8x3_t svcreate3_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_u32))) +svuint32x3_t svcreate3_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_u64))) +svuint64x3_t svcreate3_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_u16))) +svuint16x3_t svcreate3_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_s8))) +svint8x3_t svcreate3_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_f64))) +svfloat64x3_t svcreate3_f64(svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_f32))) +svfloat32x3_t svcreate3_f32(svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_f16))) +svfloat16x3_t svcreate3_f16(svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_s32))) +svint32x3_t svcreate3_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_s64))) +svint64x3_t svcreate3_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_s16))) +svint16x3_t svcreate3_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_u8))) +svuint8x4_t svcreate4_u8(svuint8_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_u32))) +svuint32x4_t svcreate4_u32(svuint32_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_u64))) +svuint64x4_t svcreate4_u64(svuint64_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_u16))) +svuint16x4_t svcreate4_u16(svuint16_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_s8))) +svint8x4_t svcreate4_s8(svint8_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_f64))) +svfloat64x4_t svcreate4_f64(svfloat64_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_f32))) +svfloat32x4_t svcreate4_f32(svfloat32_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_f16))) +svfloat16x4_t svcreate4_f16(svfloat16_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_s32))) +svint32x4_t svcreate4_s32(svint32_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_s64))) +svint64x4_t svcreate4_s64(svint64_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_s16))) +svint16x4_t svcreate4_s16(svint16_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_f32_m))) +svfloat16_t svcvt_f16_f32_m(svfloat16_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_f32_x))) +svfloat16_t svcvt_f16_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_f32_z))) +svfloat16_t svcvt_f16_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_f64_m))) +svfloat16_t svcvt_f16_f64_m(svfloat16_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_f64_x))) +svfloat16_t svcvt_f16_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_f64_z))) +svfloat16_t svcvt_f16_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s16_m))) +svfloat16_t svcvt_f16_s16_m(svfloat16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s16_x))) +svfloat16_t svcvt_f16_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s16_z))) +svfloat16_t svcvt_f16_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s32_m))) +svfloat16_t svcvt_f16_s32_m(svfloat16_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s32_x))) +svfloat16_t svcvt_f16_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s32_z))) +svfloat16_t svcvt_f16_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s64_m))) +svfloat16_t svcvt_f16_s64_m(svfloat16_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s64_x))) +svfloat16_t svcvt_f16_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s64_z))) +svfloat16_t svcvt_f16_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u16_m))) +svfloat16_t svcvt_f16_u16_m(svfloat16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u16_x))) +svfloat16_t svcvt_f16_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u16_z))) +svfloat16_t svcvt_f16_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u32_m))) +svfloat16_t svcvt_f16_u32_m(svfloat16_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u32_x))) +svfloat16_t svcvt_f16_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u32_z))) +svfloat16_t svcvt_f16_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u64_m))) +svfloat16_t svcvt_f16_u64_m(svfloat16_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u64_x))) +svfloat16_t svcvt_f16_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u64_z))) +svfloat16_t svcvt_f16_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_f16_m))) +svfloat32_t svcvt_f32_f16_m(svfloat32_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_f16_x))) +svfloat32_t svcvt_f32_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_f16_z))) +svfloat32_t svcvt_f32_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_f64_m))) +svfloat32_t svcvt_f32_f64_m(svfloat32_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_f64_x))) +svfloat32_t svcvt_f32_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_f64_z))) +svfloat32_t svcvt_f32_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_s32_m))) +svfloat32_t svcvt_f32_s32_m(svfloat32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_s32_x))) +svfloat32_t svcvt_f32_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_s32_z))) +svfloat32_t svcvt_f32_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_s64_m))) +svfloat32_t svcvt_f32_s64_m(svfloat32_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_s64_x))) +svfloat32_t svcvt_f32_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_s64_z))) +svfloat32_t svcvt_f32_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_u32_m))) +svfloat32_t svcvt_f32_u32_m(svfloat32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_u32_x))) +svfloat32_t svcvt_f32_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_u32_z))) +svfloat32_t svcvt_f32_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_u64_m))) +svfloat32_t svcvt_f32_u64_m(svfloat32_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_u64_x))) +svfloat32_t svcvt_f32_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_u64_z))) +svfloat32_t svcvt_f32_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_f16_m))) +svfloat64_t svcvt_f64_f16_m(svfloat64_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_f16_x))) +svfloat64_t svcvt_f64_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_f16_z))) +svfloat64_t svcvt_f64_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_f32_m))) +svfloat64_t svcvt_f64_f32_m(svfloat64_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_f32_x))) +svfloat64_t svcvt_f64_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_f32_z))) +svfloat64_t svcvt_f64_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_s32_m))) +svfloat64_t svcvt_f64_s32_m(svfloat64_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_s32_x))) +svfloat64_t svcvt_f64_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_s32_z))) +svfloat64_t svcvt_f64_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_s64_m))) +svfloat64_t svcvt_f64_s64_m(svfloat64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_s64_x))) +svfloat64_t svcvt_f64_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_s64_z))) +svfloat64_t svcvt_f64_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_u32_m))) +svfloat64_t svcvt_f64_u32_m(svfloat64_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_u32_x))) +svfloat64_t svcvt_f64_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_u32_z))) +svfloat64_t svcvt_f64_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_u64_m))) +svfloat64_t svcvt_f64_u64_m(svfloat64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_u64_x))) +svfloat64_t svcvt_f64_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_u64_z))) +svfloat64_t svcvt_f64_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s16_f16_m))) +svint16_t svcvt_s16_f16_m(svint16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s16_f16_x))) +svint16_t svcvt_s16_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s16_f16_z))) +svint16_t svcvt_s16_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f16_m))) +svint32_t svcvt_s32_f16_m(svint32_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f16_x))) +svint32_t svcvt_s32_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f16_z))) +svint32_t svcvt_s32_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f32_m))) +svint32_t svcvt_s32_f32_m(svint32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f32_x))) +svint32_t svcvt_s32_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f32_z))) +svint32_t svcvt_s32_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f64_m))) +svint32_t svcvt_s32_f64_m(svint32_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f64_x))) +svint32_t svcvt_s32_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f64_z))) +svint32_t svcvt_s32_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f16_m))) +svint64_t svcvt_s64_f16_m(svint64_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f16_x))) +svint64_t svcvt_s64_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f16_z))) +svint64_t svcvt_s64_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f32_m))) +svint64_t svcvt_s64_f32_m(svint64_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f32_x))) +svint64_t svcvt_s64_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f32_z))) +svint64_t svcvt_s64_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f64_m))) +svint64_t svcvt_s64_f64_m(svint64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f64_x))) +svint64_t svcvt_s64_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f64_z))) +svint64_t svcvt_s64_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u16_f16_m))) +svuint16_t svcvt_u16_f16_m(svuint16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u16_f16_x))) +svuint16_t svcvt_u16_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u16_f16_z))) +svuint16_t svcvt_u16_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f16_m))) +svuint32_t svcvt_u32_f16_m(svuint32_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f16_x))) +svuint32_t svcvt_u32_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f16_z))) +svuint32_t svcvt_u32_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f32_m))) +svuint32_t svcvt_u32_f32_m(svuint32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f32_x))) +svuint32_t svcvt_u32_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f32_z))) +svuint32_t svcvt_u32_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f64_m))) +svuint32_t svcvt_u32_f64_m(svuint32_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f64_x))) +svuint32_t svcvt_u32_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f64_z))) +svuint32_t svcvt_u32_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f16_m))) +svuint64_t svcvt_u64_f16_m(svuint64_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f16_x))) +svuint64_t svcvt_u64_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f16_z))) +svuint64_t svcvt_u64_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f32_m))) +svuint64_t svcvt_u64_f32_m(svuint64_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f32_x))) +svuint64_t svcvt_u64_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f32_z))) +svuint64_t svcvt_u64_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f64_m))) +svuint64_t svcvt_u64_f64_m(svuint64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f64_x))) +svuint64_t svcvt_u64_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f64_z))) +svuint64_t svcvt_u64_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f64_m))) +svfloat64_t svdiv_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f32_m))) +svfloat32_t svdiv_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f16_m))) +svfloat16_t svdiv_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f64_x))) +svfloat64_t svdiv_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f32_x))) +svfloat32_t svdiv_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f16_x))) +svfloat16_t svdiv_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f64_z))) +svfloat64_t svdiv_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f32_z))) +svfloat32_t svdiv_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f16_z))) +svfloat16_t svdiv_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_s32_m))) +svint32_t svdiv_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_s64_m))) +svint64_t svdiv_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_s32_x))) +svint32_t svdiv_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_s64_x))) +svint64_t svdiv_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_s32_z))) +svint32_t svdiv_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_s64_z))) +svint64_t svdiv_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_u32_m))) +svuint32_t svdiv_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_u64_m))) +svuint64_t svdiv_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_u32_x))) +svuint32_t svdiv_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_u64_x))) +svuint64_t svdiv_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_u32_z))) +svuint32_t svdiv_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_u64_z))) +svuint64_t svdiv_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f64_m))) +svfloat64_t svdiv_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f32_m))) +svfloat32_t svdiv_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f16_m))) +svfloat16_t svdiv_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f64_x))) +svfloat64_t svdiv_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f32_x))) +svfloat32_t svdiv_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f16_x))) +svfloat16_t svdiv_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f64_z))) +svfloat64_t svdiv_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f32_z))) +svfloat32_t svdiv_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f16_z))) +svfloat16_t svdiv_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_s32_m))) +svint32_t svdiv_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_s64_m))) +svint64_t svdiv_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_s32_x))) +svint32_t svdiv_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_s64_x))) +svint64_t svdiv_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_s32_z))) +svint32_t svdiv_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_s64_z))) +svint64_t svdiv_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_u32_m))) +svuint32_t svdiv_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_u64_m))) +svuint64_t svdiv_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_u32_x))) +svuint32_t svdiv_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_u64_x))) +svuint64_t svdiv_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_u32_z))) +svuint32_t svdiv_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_u64_z))) +svuint64_t svdiv_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f64_m))) +svfloat64_t svdivr_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f32_m))) +svfloat32_t svdivr_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f16_m))) +svfloat16_t svdivr_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f64_x))) +svfloat64_t svdivr_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f32_x))) +svfloat32_t svdivr_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f16_x))) +svfloat16_t svdivr_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f64_z))) +svfloat64_t svdivr_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f32_z))) +svfloat32_t svdivr_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f16_z))) +svfloat16_t svdivr_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_s32_m))) +svint32_t svdivr_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_s64_m))) +svint64_t svdivr_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_s32_x))) +svint32_t svdivr_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_s64_x))) +svint64_t svdivr_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_s32_z))) +svint32_t svdivr_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_s64_z))) +svint64_t svdivr_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_u32_m))) +svuint32_t svdivr_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_u64_m))) +svuint64_t svdivr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_u32_x))) +svuint32_t svdivr_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_u64_x))) +svuint64_t svdivr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_u32_z))) +svuint32_t svdivr_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_u64_z))) +svuint64_t svdivr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f64_m))) +svfloat64_t svdivr_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f32_m))) +svfloat32_t svdivr_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f16_m))) +svfloat16_t svdivr_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f64_x))) +svfloat64_t svdivr_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f32_x))) +svfloat32_t svdivr_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f16_x))) +svfloat16_t svdivr_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f64_z))) +svfloat64_t svdivr_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f32_z))) +svfloat32_t svdivr_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f16_z))) +svfloat16_t svdivr_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_s32_m))) +svint32_t svdivr_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_s64_m))) +svint64_t svdivr_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_s32_x))) +svint32_t svdivr_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_s64_x))) +svint64_t svdivr_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_s32_z))) +svint32_t svdivr_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_s64_z))) +svint64_t svdivr_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_u32_m))) +svuint32_t svdivr_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_u64_m))) +svuint64_t svdivr_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_u32_x))) +svuint32_t svdivr_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_u64_x))) +svuint64_t svdivr_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_u32_z))) +svuint32_t svdivr_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_u64_z))) +svuint64_t svdivr_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_n_s32))) +svint32_t svdot_n_s32(svint32_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_n_s64))) +svint64_t svdot_n_s64(svint64_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_n_u32))) +svuint32_t svdot_n_u32(svuint32_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_n_u64))) +svuint64_t svdot_n_u64(svuint64_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_s32))) +svint32_t svdot_s32(svint32_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_s64))) +svint64_t svdot_s64(svint64_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_u32))) +svuint32_t svdot_u32(svuint32_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_u64))) +svuint64_t svdot_u64(svuint64_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_lane_s32))) +svint32_t svdot_lane_s32(svint32_t, svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_lane_s64))) +svint64_t svdot_lane_s64(svint64_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_lane_u32))) +svuint32_t svdot_lane_u32(svuint32_t, svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_lane_u64))) +svuint64_t svdot_lane_u64(svuint64_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u8))) +svuint8_t svdup_n_u8(uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u32))) +svuint32_t svdup_n_u32(uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u64))) +svuint64_t svdup_n_u64(uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u16))) +svuint16_t svdup_n_u16(uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s8))) +svint8_t svdup_n_s8(int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f64))) +svfloat64_t svdup_n_f64(float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f32))) +svfloat32_t svdup_n_f32(float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f16))) +svfloat16_t svdup_n_f16(float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s32))) +svint32_t svdup_n_s32(int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s64))) +svint64_t svdup_n_s64(int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s16))) +svint16_t svdup_n_s16(int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u8_m))) +svuint8_t svdup_n_u8_m(svuint8_t, svbool_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u32_m))) +svuint32_t svdup_n_u32_m(svuint32_t, svbool_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u64_m))) +svuint64_t svdup_n_u64_m(svuint64_t, svbool_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u16_m))) +svuint16_t svdup_n_u16_m(svuint16_t, svbool_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s8_m))) +svint8_t svdup_n_s8_m(svint8_t, svbool_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f64_m))) +svfloat64_t svdup_n_f64_m(svfloat64_t, svbool_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f32_m))) +svfloat32_t svdup_n_f32_m(svfloat32_t, svbool_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f16_m))) +svfloat16_t svdup_n_f16_m(svfloat16_t, svbool_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s32_m))) +svint32_t svdup_n_s32_m(svint32_t, svbool_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s64_m))) +svint64_t svdup_n_s64_m(svint64_t, svbool_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s16_m))) +svint16_t svdup_n_s16_m(svint16_t, svbool_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_b8))) +svbool_t svdup_n_b8(bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_b32))) +svbool_t svdup_n_b32(bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_b64))) +svbool_t svdup_n_b64(bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_b16))) +svbool_t svdup_n_b16(bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u8_x))) +svuint8_t svdup_n_u8_x(svbool_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u32_x))) +svuint32_t svdup_n_u32_x(svbool_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u64_x))) +svuint64_t svdup_n_u64_x(svbool_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u16_x))) +svuint16_t svdup_n_u16_x(svbool_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s8_x))) +svint8_t svdup_n_s8_x(svbool_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f64_x))) +svfloat64_t svdup_n_f64_x(svbool_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f32_x))) +svfloat32_t svdup_n_f32_x(svbool_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f16_x))) +svfloat16_t svdup_n_f16_x(svbool_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s32_x))) +svint32_t svdup_n_s32_x(svbool_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s64_x))) +svint64_t svdup_n_s64_x(svbool_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s16_x))) +svint16_t svdup_n_s16_x(svbool_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u8_z))) +svuint8_t svdup_n_u8_z(svbool_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u32_z))) +svuint32_t svdup_n_u32_z(svbool_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u64_z))) +svuint64_t svdup_n_u64_z(svbool_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u16_z))) +svuint16_t svdup_n_u16_z(svbool_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s8_z))) +svint8_t svdup_n_s8_z(svbool_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f64_z))) +svfloat64_t svdup_n_f64_z(svbool_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f32_z))) +svfloat32_t svdup_n_f32_z(svbool_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f16_z))) +svfloat16_t svdup_n_f16_z(svbool_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s32_z))) +svint32_t svdup_n_s32_z(svbool_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s64_z))) +svint64_t svdup_n_s64_z(svbool_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s16_z))) +svint16_t svdup_n_s16_z(svbool_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_u8))) +svuint8_t svdup_lane_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_u32))) +svuint32_t svdup_lane_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_u64))) +svuint64_t svdup_lane_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_u16))) +svuint16_t svdup_lane_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_s8))) +svint8_t svdup_lane_s8(svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_f64))) +svfloat64_t svdup_lane_f64(svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_f32))) +svfloat32_t svdup_lane_f32(svfloat32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_f16))) +svfloat16_t svdup_lane_f16(svfloat16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_s32))) +svint32_t svdup_lane_s32(svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_s64))) +svint64_t svdup_lane_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_s16))) +svint16_t svdup_lane_s16(svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_u16))) +svuint16_t svdupq_n_u16(uint16_t, uint16_t, uint16_t, uint16_t, uint16_t, uint16_t, uint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_f16))) +svfloat16_t svdupq_n_f16(float16_t, float16_t, float16_t, float16_t, float16_t, float16_t, float16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_s16))) +svint16_t svdupq_n_s16(int16_t, int16_t, int16_t, int16_t, int16_t, int16_t, int16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_u32))) +svuint32_t svdupq_n_u32(uint32_t, uint32_t, uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_f32))) +svfloat32_t svdupq_n_f32(float32_t, float32_t, float32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_s32))) +svint32_t svdupq_n_s32(int32_t, int32_t, int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_u64))) +svuint64_t svdupq_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_f64))) +svfloat64_t svdupq_n_f64(float64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_s64))) +svint64_t svdupq_n_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_u8))) +svuint8_t svdupq_n_u8(uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_s8))) +svint8_t svdupq_n_s8(int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_b16))) +svbool_t svdupq_n_b16(bool, bool, bool, bool, bool, bool, bool, bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_b32))) +svbool_t svdupq_n_b32(bool, bool, bool, bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_b64))) +svbool_t svdupq_n_b64(bool, bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_b8))) +svbool_t svdupq_n_b8(bool, bool, bool, bool, bool, bool, bool, bool, bool, bool, bool, bool, bool, bool, bool, bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_u8))) +svuint8_t svdupq_lane_u8(svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_u32))) +svuint32_t svdupq_lane_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_u64))) +svuint64_t svdupq_lane_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_u16))) +svuint16_t svdupq_lane_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_s8))) +svint8_t svdupq_lane_s8(svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_f64))) +svfloat64_t svdupq_lane_f64(svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_f32))) +svfloat32_t svdupq_lane_f32(svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_f16))) +svfloat16_t svdupq_lane_f16(svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_s32))) +svint32_t svdupq_lane_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_s64))) +svint64_t svdupq_lane_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_s16))) +svint16_t svdupq_lane_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_b_z))) +svbool_t sveor_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u8_m))) +svuint8_t sveor_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u32_m))) +svuint32_t sveor_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u64_m))) +svuint64_t sveor_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u16_m))) +svuint16_t sveor_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s8_m))) +svint8_t sveor_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s32_m))) +svint32_t sveor_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s64_m))) +svint64_t sveor_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s16_m))) +svint16_t sveor_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u8_x))) +svuint8_t sveor_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u32_x))) +svuint32_t sveor_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u64_x))) +svuint64_t sveor_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u16_x))) +svuint16_t sveor_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s8_x))) +svint8_t sveor_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s32_x))) +svint32_t sveor_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s64_x))) +svint64_t sveor_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s16_x))) +svint16_t sveor_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u8_z))) +svuint8_t sveor_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u32_z))) +svuint32_t sveor_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u64_z))) +svuint64_t sveor_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u16_z))) +svuint16_t sveor_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s8_z))) +svint8_t sveor_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s32_z))) +svint32_t sveor_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s64_z))) +svint64_t sveor_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s16_z))) +svint16_t sveor_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u8_m))) +svuint8_t sveor_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u32_m))) +svuint32_t sveor_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u64_m))) +svuint64_t sveor_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u16_m))) +svuint16_t sveor_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s8_m))) +svint8_t sveor_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s32_m))) +svint32_t sveor_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s64_m))) +svint64_t sveor_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s16_m))) +svint16_t sveor_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u8_x))) +svuint8_t sveor_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u32_x))) +svuint32_t sveor_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u64_x))) +svuint64_t sveor_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u16_x))) +svuint16_t sveor_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s8_x))) +svint8_t sveor_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s32_x))) +svint32_t sveor_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s64_x))) +svint64_t sveor_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s16_x))) +svint16_t sveor_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u8_z))) +svuint8_t sveor_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u32_z))) +svuint32_t sveor_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u64_z))) +svuint64_t sveor_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u16_z))) +svuint16_t sveor_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s8_z))) +svint8_t sveor_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s32_z))) +svint32_t sveor_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s64_z))) +svint64_t sveor_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s16_z))) +svint16_t sveor_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_u8))) +uint8_t sveorv_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_u32))) +uint32_t sveorv_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_u64))) +uint64_t sveorv_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_u16))) +uint16_t sveorv_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_s8))) +int8_t sveorv_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_s32))) +int32_t sveorv_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_s64))) +int64_t sveorv_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_s16))) +int16_t sveorv_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexpa_f64))) +svfloat64_t svexpa_f64(svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexpa_f32))) +svfloat32_t svexpa_f32(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexpa_f16))) +svfloat16_t svexpa_f16(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_u8))) +svuint8_t svext_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_u32))) +svuint32_t svext_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_u64))) +svuint64_t svext_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_u16))) +svuint16_t svext_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_s8))) +svint8_t svext_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_f64))) +svfloat64_t svext_f64(svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_f32))) +svfloat32_t svext_f32(svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_f16))) +svfloat16_t svext_f16(svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_s32))) +svint32_t svext_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_s64))) +svint64_t svext_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_s16))) +svint16_t svext_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s32_m))) +svint32_t svextb_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s64_m))) +svint64_t svextb_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s16_m))) +svint16_t svextb_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s32_x))) +svint32_t svextb_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s64_x))) +svint64_t svextb_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s16_x))) +svint16_t svextb_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s32_z))) +svint32_t svextb_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s64_z))) +svint64_t svextb_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s16_z))) +svint16_t svextb_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u32_m))) +svuint32_t svextb_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u64_m))) +svuint64_t svextb_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u16_m))) +svuint16_t svextb_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u32_x))) +svuint32_t svextb_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u64_x))) +svuint64_t svextb_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u16_x))) +svuint16_t svextb_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u32_z))) +svuint32_t svextb_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u64_z))) +svuint64_t svextb_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u16_z))) +svuint16_t svextb_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_s32_m))) +svint32_t svexth_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_s64_m))) +svint64_t svexth_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_s32_x))) +svint32_t svexth_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_s64_x))) +svint64_t svexth_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_s32_z))) +svint32_t svexth_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_s64_z))) +svint64_t svexth_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_u32_m))) +svuint32_t svexth_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_u64_m))) +svuint64_t svexth_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_u32_x))) +svuint32_t svexth_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_u64_x))) +svuint64_t svexth_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_u32_z))) +svuint32_t svexth_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_u64_z))) +svuint64_t svexth_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextw_s64_m))) +svint64_t svextw_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextw_s64_x))) +svint64_t svextw_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextw_s64_z))) +svint64_t svextw_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextw_u64_m))) +svuint64_t svextw_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextw_u64_x))) +svuint64_t svextw_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextw_u64_z))) +svuint64_t svextw_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_u8))) +svuint8_t svget2_u8(svuint8x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_u32))) +svuint32_t svget2_u32(svuint32x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_u64))) +svuint64_t svget2_u64(svuint64x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_u16))) +svuint16_t svget2_u16(svuint16x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_s8))) +svint8_t svget2_s8(svint8x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_f64))) +svfloat64_t svget2_f64(svfloat64x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_f32))) +svfloat32_t svget2_f32(svfloat32x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_f16))) +svfloat16_t svget2_f16(svfloat16x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_s32))) +svint32_t svget2_s32(svint32x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_s64))) +svint64_t svget2_s64(svint64x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_s16))) +svint16_t svget2_s16(svint16x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_u8))) +svuint8_t svget3_u8(svuint8x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_u32))) +svuint32_t svget3_u32(svuint32x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_u64))) +svuint64_t svget3_u64(svuint64x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_u16))) +svuint16_t svget3_u16(svuint16x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_s8))) +svint8_t svget3_s8(svint8x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_f64))) +svfloat64_t svget3_f64(svfloat64x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_f32))) +svfloat32_t svget3_f32(svfloat32x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_f16))) +svfloat16_t svget3_f16(svfloat16x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_s32))) +svint32_t svget3_s32(svint32x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_s64))) +svint64_t svget3_s64(svint64x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_s16))) +svint16_t svget3_s16(svint16x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_u8))) +svuint8_t svget4_u8(svuint8x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_u32))) +svuint32_t svget4_u32(svuint32x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_u64))) +svuint64_t svget4_u64(svuint64x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_u16))) +svuint16_t svget4_u16(svuint16x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_s8))) +svint8_t svget4_s8(svint8x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_f64))) +svfloat64_t svget4_f64(svfloat64x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_f32))) +svfloat32_t svget4_f32(svfloat32x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_f16))) +svfloat16_t svget4_f16(svfloat16x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_s32))) +svint32_t svget4_s32(svint32x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_s64))) +svint64_t svget4_s64(svint64x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_s16))) +svint16_t svget4_s16(svint16x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_u8))) +svuint8_t svindex_u8(uint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_u32))) +svuint32_t svindex_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_u64))) +svuint64_t svindex_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_u16))) +svuint16_t svindex_u16(uint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_s8))) +svint8_t svindex_s8(int8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_s32))) +svint32_t svindex_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_s64))) +svint64_t svindex_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_s16))) +svint16_t svindex_s16(int16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_u8))) +svuint8_t svinsr_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_u32))) +svuint32_t svinsr_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_u64))) +svuint64_t svinsr_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_u16))) +svuint16_t svinsr_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_s8))) +svint8_t svinsr_n_s8(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_f64))) +svfloat64_t svinsr_n_f64(svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_f32))) +svfloat32_t svinsr_n_f32(svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_f16))) +svfloat16_t svinsr_n_f16(svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_s32))) +svint32_t svinsr_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_s64))) +svint64_t svinsr_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_s16))) +svint16_t svinsr_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_u8))) +uint8_t svlasta_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_u32))) +uint32_t svlasta_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_u64))) +uint64_t svlasta_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_u16))) +uint16_t svlasta_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_s8))) +int8_t svlasta_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_f64))) +float64_t svlasta_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_f32))) +float32_t svlasta_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_f16))) +float16_t svlasta_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_s32))) +int32_t svlasta_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_s64))) +int64_t svlasta_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_s16))) +int16_t svlasta_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_u8))) +uint8_t svlastb_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_u32))) +uint32_t svlastb_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_u64))) +uint64_t svlastb_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_u16))) +uint16_t svlastb_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_s8))) +int8_t svlastb_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_f64))) +float64_t svlastb_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_f32))) +float32_t svlastb_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_f16))) +float16_t svlastb_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_s32))) +int32_t svlastb_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_s64))) +int64_t svlastb_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_s16))) +int16_t svlastb_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_u8))) +svuint8_t svld1_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_u32))) +svuint32_t svld1_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_u64))) +svuint64_t svld1_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_u16))) +svuint16_t svld1_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_s8))) +svint8_t svld1_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_f64))) +svfloat64_t svld1_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_f32))) +svfloat32_t svld1_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_f16))) +svfloat16_t svld1_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_s32))) +svint32_t svld1_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_s64))) +svint64_t svld1_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_s16))) +svint16_t svld1_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_index_u32))) +svuint32_t svld1_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_index_u64))) +svuint64_t svld1_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_index_f64))) +svfloat64_t svld1_gather_u64base_index_f64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_index_f32))) +svfloat32_t svld1_gather_u32base_index_f32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_index_s32))) +svint32_t svld1_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_index_s64))) +svint64_t svld1_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_offset_u32))) +svuint32_t svld1_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_offset_u64))) +svuint64_t svld1_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_offset_f64))) +svfloat64_t svld1_gather_u64base_offset_f64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_offset_f32))) +svfloat32_t svld1_gather_u32base_offset_f32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_offset_s32))) +svint32_t svld1_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_offset_s64))) +svint64_t svld1_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_u32))) +svuint32_t svld1_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_u64))) +svuint64_t svld1_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_f64))) +svfloat64_t svld1_gather_u64base_f64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_f32))) +svfloat32_t svld1_gather_u32base_f32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_s32))) +svint32_t svld1_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_s64))) +svint64_t svld1_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s32index_u32))) +svuint32_t svld1_gather_s32index_u32(svbool_t, uint32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s32index_f32))) +svfloat32_t svld1_gather_s32index_f32(svbool_t, float32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s32index_s32))) +svint32_t svld1_gather_s32index_s32(svbool_t, int32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32index_u32))) +svuint32_t svld1_gather_u32index_u32(svbool_t, uint32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32index_f32))) +svfloat32_t svld1_gather_u32index_f32(svbool_t, float32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32index_s32))) +svint32_t svld1_gather_u32index_s32(svbool_t, int32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s64index_u64))) +svuint64_t svld1_gather_s64index_u64(svbool_t, uint64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s64index_f64))) +svfloat64_t svld1_gather_s64index_f64(svbool_t, float64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s64index_s64))) +svint64_t svld1_gather_s64index_s64(svbool_t, int64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64index_u64))) +svuint64_t svld1_gather_u64index_u64(svbool_t, uint64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64index_f64))) +svfloat64_t svld1_gather_u64index_f64(svbool_t, float64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64index_s64))) +svint64_t svld1_gather_u64index_s64(svbool_t, int64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s32offset_u32))) +svuint32_t svld1_gather_s32offset_u32(svbool_t, uint32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s32offset_f32))) +svfloat32_t svld1_gather_s32offset_f32(svbool_t, float32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s32offset_s32))) +svint32_t svld1_gather_s32offset_s32(svbool_t, int32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32offset_u32))) +svuint32_t svld1_gather_u32offset_u32(svbool_t, uint32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32offset_f32))) +svfloat32_t svld1_gather_u32offset_f32(svbool_t, float32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32offset_s32))) +svint32_t svld1_gather_u32offset_s32(svbool_t, int32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s64offset_u64))) +svuint64_t svld1_gather_s64offset_u64(svbool_t, uint64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s64offset_f64))) +svfloat64_t svld1_gather_s64offset_f64(svbool_t, float64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s64offset_s64))) +svint64_t svld1_gather_s64offset_s64(svbool_t, int64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64offset_u64))) +svuint64_t svld1_gather_u64offset_u64(svbool_t, uint64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64offset_f64))) +svfloat64_t svld1_gather_u64offset_f64(svbool_t, float64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64offset_s64))) +svint64_t svld1_gather_u64offset_s64(svbool_t, int64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_u8))) +svuint8_t svld1_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_u32))) +svuint32_t svld1_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_u64))) +svuint64_t svld1_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_u16))) +svuint16_t svld1_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_s8))) +svint8_t svld1_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_f64))) +svfloat64_t svld1_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_f32))) +svfloat32_t svld1_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_f16))) +svfloat16_t svld1_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_s32))) +svint32_t svld1_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_s64))) +svint64_t svld1_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_s16))) +svint16_t svld1_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_u8))) +svuint8_t svld1rq_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_u32))) +svuint32_t svld1rq_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_u64))) +svuint64_t svld1rq_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_u16))) +svuint16_t svld1rq_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_s8))) +svint8_t svld1rq_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_f64))) +svfloat64_t svld1rq_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_f32))) +svfloat32_t svld1rq_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_f16))) +svfloat16_t svld1rq_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_s32))) +svint32_t svld1rq_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_s64))) +svint64_t svld1rq_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_s16))) +svint16_t svld1rq_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u32base_offset_u32))) +svuint32_t svld1sb_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u64base_offset_u64))) +svuint64_t svld1sb_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u32base_offset_s32))) +svint32_t svld1sb_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u64base_offset_s64))) +svint64_t svld1sb_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u32base_u32))) +svuint32_t svld1sb_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u64base_u64))) +svuint64_t svld1sb_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u32base_s32))) +svint32_t svld1sb_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u64base_s64))) +svint64_t svld1sb_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_s32offset_u32))) +svuint32_t svld1sb_gather_s32offset_u32(svbool_t, int8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_s32offset_s32))) +svint32_t svld1sb_gather_s32offset_s32(svbool_t, int8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u32offset_u32))) +svuint32_t svld1sb_gather_u32offset_u32(svbool_t, int8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u32offset_s32))) +svint32_t svld1sb_gather_u32offset_s32(svbool_t, int8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_s64offset_u64))) +svuint64_t svld1sb_gather_s64offset_u64(svbool_t, int8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_s64offset_s64))) +svint64_t svld1sb_gather_s64offset_s64(svbool_t, int8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u64offset_u64))) +svuint64_t svld1sb_gather_u64offset_u64(svbool_t, int8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u64offset_s64))) +svint64_t svld1sb_gather_u64offset_s64(svbool_t, int8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_vnum_u32))) +svuint32_t svld1sb_vnum_u32(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_vnum_u64))) +svuint64_t svld1sb_vnum_u64(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_vnum_u16))) +svuint16_t svld1sb_vnum_u16(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_vnum_s32))) +svint32_t svld1sb_vnum_s32(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_vnum_s64))) +svint64_t svld1sb_vnum_s64(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_vnum_s16))) +svint16_t svld1sb_vnum_s16(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_u32))) +svuint32_t svld1sb_u32(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_u64))) +svuint64_t svld1sb_u64(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_u16))) +svuint16_t svld1sb_u16(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_s32))) +svint32_t svld1sb_s32(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_s64))) +svint64_t svld1sb_s64(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_s16))) +svint16_t svld1sb_s16(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32base_index_u32))) +svuint32_t svld1sh_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64base_index_u64))) +svuint64_t svld1sh_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32base_index_s32))) +svint32_t svld1sh_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64base_index_s64))) +svint64_t svld1sh_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32base_offset_u32))) +svuint32_t svld1sh_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64base_offset_u64))) +svuint64_t svld1sh_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32base_offset_s32))) +svint32_t svld1sh_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64base_offset_s64))) +svint64_t svld1sh_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32base_u32))) +svuint32_t svld1sh_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64base_u64))) +svuint64_t svld1sh_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32base_s32))) +svint32_t svld1sh_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64base_s64))) +svint64_t svld1sh_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s32index_u32))) +svuint32_t svld1sh_gather_s32index_u32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s32index_s32))) +svint32_t svld1sh_gather_s32index_s32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32index_u32))) +svuint32_t svld1sh_gather_u32index_u32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32index_s32))) +svint32_t svld1sh_gather_u32index_s32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s64index_u64))) +svuint64_t svld1sh_gather_s64index_u64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s64index_s64))) +svint64_t svld1sh_gather_s64index_s64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64index_u64))) +svuint64_t svld1sh_gather_u64index_u64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64index_s64))) +svint64_t svld1sh_gather_u64index_s64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s32offset_u32))) +svuint32_t svld1sh_gather_s32offset_u32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s32offset_s32))) +svint32_t svld1sh_gather_s32offset_s32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32offset_u32))) +svuint32_t svld1sh_gather_u32offset_u32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32offset_s32))) +svint32_t svld1sh_gather_u32offset_s32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s64offset_u64))) +svuint64_t svld1sh_gather_s64offset_u64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s64offset_s64))) +svint64_t svld1sh_gather_s64offset_s64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64offset_u64))) +svuint64_t svld1sh_gather_u64offset_u64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64offset_s64))) +svint64_t svld1sh_gather_u64offset_s64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_vnum_u32))) +svuint32_t svld1sh_vnum_u32(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_vnum_u64))) +svuint64_t svld1sh_vnum_u64(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_vnum_s32))) +svint32_t svld1sh_vnum_s32(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_vnum_s64))) +svint64_t svld1sh_vnum_s64(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_u32))) +svuint32_t svld1sh_u32(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_u64))) +svuint64_t svld1sh_u64(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_s32))) +svint32_t svld1sh_s32(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_s64))) +svint64_t svld1sh_s64(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64base_index_u64))) +svuint64_t svld1sw_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64base_index_s64))) +svint64_t svld1sw_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64base_offset_u64))) +svuint64_t svld1sw_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64base_offset_s64))) +svint64_t svld1sw_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64base_u64))) +svuint64_t svld1sw_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64base_s64))) +svint64_t svld1sw_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_s64index_u64))) +svuint64_t svld1sw_gather_s64index_u64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_s64index_s64))) +svint64_t svld1sw_gather_s64index_s64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64index_u64))) +svuint64_t svld1sw_gather_u64index_u64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64index_s64))) +svint64_t svld1sw_gather_u64index_s64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_s64offset_u64))) +svuint64_t svld1sw_gather_s64offset_u64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_s64offset_s64))) +svint64_t svld1sw_gather_s64offset_s64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64offset_u64))) +svuint64_t svld1sw_gather_u64offset_u64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64offset_s64))) +svint64_t svld1sw_gather_u64offset_s64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_vnum_u64))) +svuint64_t svld1sw_vnum_u64(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_vnum_s64))) +svint64_t svld1sw_vnum_s64(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_u64))) +svuint64_t svld1sw_u64(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_s64))) +svint64_t svld1sw_s64(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u32base_offset_u32))) +svuint32_t svld1ub_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u64base_offset_u64))) +svuint64_t svld1ub_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u32base_offset_s32))) +svint32_t svld1ub_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u64base_offset_s64))) +svint64_t svld1ub_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u32base_u32))) +svuint32_t svld1ub_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u64base_u64))) +svuint64_t svld1ub_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u32base_s32))) +svint32_t svld1ub_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u64base_s64))) +svint64_t svld1ub_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_s32offset_u32))) +svuint32_t svld1ub_gather_s32offset_u32(svbool_t, uint8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_s32offset_s32))) +svint32_t svld1ub_gather_s32offset_s32(svbool_t, uint8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u32offset_u32))) +svuint32_t svld1ub_gather_u32offset_u32(svbool_t, uint8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u32offset_s32))) +svint32_t svld1ub_gather_u32offset_s32(svbool_t, uint8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_s64offset_u64))) +svuint64_t svld1ub_gather_s64offset_u64(svbool_t, uint8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_s64offset_s64))) +svint64_t svld1ub_gather_s64offset_s64(svbool_t, uint8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u64offset_u64))) +svuint64_t svld1ub_gather_u64offset_u64(svbool_t, uint8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u64offset_s64))) +svint64_t svld1ub_gather_u64offset_s64(svbool_t, uint8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_vnum_u32))) +svuint32_t svld1ub_vnum_u32(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_vnum_u64))) +svuint64_t svld1ub_vnum_u64(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_vnum_u16))) +svuint16_t svld1ub_vnum_u16(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_vnum_s32))) +svint32_t svld1ub_vnum_s32(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_vnum_s64))) +svint64_t svld1ub_vnum_s64(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_vnum_s16))) +svint16_t svld1ub_vnum_s16(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_u32))) +svuint32_t svld1ub_u32(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_u64))) +svuint64_t svld1ub_u64(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_u16))) +svuint16_t svld1ub_u16(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_s32))) +svint32_t svld1ub_s32(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_s64))) +svint64_t svld1ub_s64(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_s16))) +svint16_t svld1ub_s16(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32base_index_u32))) +svuint32_t svld1uh_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64base_index_u64))) +svuint64_t svld1uh_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32base_index_s32))) +svint32_t svld1uh_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64base_index_s64))) +svint64_t svld1uh_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32base_offset_u32))) +svuint32_t svld1uh_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64base_offset_u64))) +svuint64_t svld1uh_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32base_offset_s32))) +svint32_t svld1uh_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64base_offset_s64))) +svint64_t svld1uh_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32base_u32))) +svuint32_t svld1uh_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64base_u64))) +svuint64_t svld1uh_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32base_s32))) +svint32_t svld1uh_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64base_s64))) +svint64_t svld1uh_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s32index_u32))) +svuint32_t svld1uh_gather_s32index_u32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s32index_s32))) +svint32_t svld1uh_gather_s32index_s32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32index_u32))) +svuint32_t svld1uh_gather_u32index_u32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32index_s32))) +svint32_t svld1uh_gather_u32index_s32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s64index_u64))) +svuint64_t svld1uh_gather_s64index_u64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s64index_s64))) +svint64_t svld1uh_gather_s64index_s64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64index_u64))) +svuint64_t svld1uh_gather_u64index_u64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64index_s64))) +svint64_t svld1uh_gather_u64index_s64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s32offset_u32))) +svuint32_t svld1uh_gather_s32offset_u32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s32offset_s32))) +svint32_t svld1uh_gather_s32offset_s32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32offset_u32))) +svuint32_t svld1uh_gather_u32offset_u32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32offset_s32))) +svint32_t svld1uh_gather_u32offset_s32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s64offset_u64))) +svuint64_t svld1uh_gather_s64offset_u64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s64offset_s64))) +svint64_t svld1uh_gather_s64offset_s64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64offset_u64))) +svuint64_t svld1uh_gather_u64offset_u64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64offset_s64))) +svint64_t svld1uh_gather_u64offset_s64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_vnum_u32))) +svuint32_t svld1uh_vnum_u32(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_vnum_u64))) +svuint64_t svld1uh_vnum_u64(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_vnum_s32))) +svint32_t svld1uh_vnum_s32(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_vnum_s64))) +svint64_t svld1uh_vnum_s64(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_u32))) +svuint32_t svld1uh_u32(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_u64))) +svuint64_t svld1uh_u64(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_s32))) +svint32_t svld1uh_s32(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_s64))) +svint64_t svld1uh_s64(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64base_index_u64))) +svuint64_t svld1uw_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64base_index_s64))) +svint64_t svld1uw_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64base_offset_u64))) +svuint64_t svld1uw_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64base_offset_s64))) +svint64_t svld1uw_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64base_u64))) +svuint64_t svld1uw_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64base_s64))) +svint64_t svld1uw_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_s64index_u64))) +svuint64_t svld1uw_gather_s64index_u64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_s64index_s64))) +svint64_t svld1uw_gather_s64index_s64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64index_u64))) +svuint64_t svld1uw_gather_u64index_u64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64index_s64))) +svint64_t svld1uw_gather_u64index_s64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_s64offset_u64))) +svuint64_t svld1uw_gather_s64offset_u64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_s64offset_s64))) +svint64_t svld1uw_gather_s64offset_s64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64offset_u64))) +svuint64_t svld1uw_gather_u64offset_u64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64offset_s64))) +svint64_t svld1uw_gather_u64offset_s64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_vnum_u64))) +svuint64_t svld1uw_vnum_u64(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_vnum_s64))) +svint64_t svld1uw_vnum_s64(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_u64))) +svuint64_t svld1uw_u64(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_s64))) +svint64_t svld1uw_s64(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_u8))) +svuint8x2_t svld2_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_u32))) +svuint32x2_t svld2_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_u64))) +svuint64x2_t svld2_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_u16))) +svuint16x2_t svld2_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_s8))) +svint8x2_t svld2_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_f64))) +svfloat64x2_t svld2_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_f32))) +svfloat32x2_t svld2_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_f16))) +svfloat16x2_t svld2_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_s32))) +svint32x2_t svld2_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_s64))) +svint64x2_t svld2_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_s16))) +svint16x2_t svld2_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_u8))) +svuint8x2_t svld2_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_u32))) +svuint32x2_t svld2_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_u64))) +svuint64x2_t svld2_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_u16))) +svuint16x2_t svld2_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_s8))) +svint8x2_t svld2_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_f64))) +svfloat64x2_t svld2_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_f32))) +svfloat32x2_t svld2_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_f16))) +svfloat16x2_t svld2_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_s32))) +svint32x2_t svld2_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_s64))) +svint64x2_t svld2_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_s16))) +svint16x2_t svld2_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_u8))) +svuint8x3_t svld3_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_u32))) +svuint32x3_t svld3_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_u64))) +svuint64x3_t svld3_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_u16))) +svuint16x3_t svld3_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_s8))) +svint8x3_t svld3_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_f64))) +svfloat64x3_t svld3_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_f32))) +svfloat32x3_t svld3_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_f16))) +svfloat16x3_t svld3_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_s32))) +svint32x3_t svld3_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_s64))) +svint64x3_t svld3_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_s16))) +svint16x3_t svld3_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_u8))) +svuint8x3_t svld3_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_u32))) +svuint32x3_t svld3_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_u64))) +svuint64x3_t svld3_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_u16))) +svuint16x3_t svld3_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_s8))) +svint8x3_t svld3_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_f64))) +svfloat64x3_t svld3_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_f32))) +svfloat32x3_t svld3_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_f16))) +svfloat16x3_t svld3_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_s32))) +svint32x3_t svld3_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_s64))) +svint64x3_t svld3_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_s16))) +svint16x3_t svld3_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_u8))) +svuint8x4_t svld4_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_u32))) +svuint32x4_t svld4_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_u64))) +svuint64x4_t svld4_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_u16))) +svuint16x4_t svld4_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_s8))) +svint8x4_t svld4_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_f64))) +svfloat64x4_t svld4_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_f32))) +svfloat32x4_t svld4_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_f16))) +svfloat16x4_t svld4_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_s32))) +svint32x4_t svld4_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_s64))) +svint64x4_t svld4_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_s16))) +svint16x4_t svld4_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_u8))) +svuint8x4_t svld4_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_u32))) +svuint32x4_t svld4_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_u64))) +svuint64x4_t svld4_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_u16))) +svuint16x4_t svld4_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_s8))) +svint8x4_t svld4_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_f64))) +svfloat64x4_t svld4_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_f32))) +svfloat32x4_t svld4_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_f16))) +svfloat16x4_t svld4_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_s32))) +svint32x4_t svld4_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_s64))) +svint64x4_t svld4_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_s16))) +svint16x4_t svld4_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_u8))) +svuint8_t svldff1_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_u32))) +svuint32_t svldff1_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_u64))) +svuint64_t svldff1_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_u16))) +svuint16_t svldff1_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_s8))) +svint8_t svldff1_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_f64))) +svfloat64_t svldff1_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_f32))) +svfloat32_t svldff1_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_f16))) +svfloat16_t svldff1_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_s32))) +svint32_t svldff1_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_s64))) +svint64_t svldff1_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_s16))) +svint16_t svldff1_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_index_u32))) +svuint32_t svldff1_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_index_u64))) +svuint64_t svldff1_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_index_f64))) +svfloat64_t svldff1_gather_u64base_index_f64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_index_f32))) +svfloat32_t svldff1_gather_u32base_index_f32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_index_s32))) +svint32_t svldff1_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_index_s64))) +svint64_t svldff1_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_offset_u32))) +svuint32_t svldff1_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_offset_u64))) +svuint64_t svldff1_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_offset_f64))) +svfloat64_t svldff1_gather_u64base_offset_f64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_offset_f32))) +svfloat32_t svldff1_gather_u32base_offset_f32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_offset_s32))) +svint32_t svldff1_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_offset_s64))) +svint64_t svldff1_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_u32))) +svuint32_t svldff1_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_u64))) +svuint64_t svldff1_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_f64))) +svfloat64_t svldff1_gather_u64base_f64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_f32))) +svfloat32_t svldff1_gather_u32base_f32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_s32))) +svint32_t svldff1_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_s64))) +svint64_t svldff1_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s32index_u32))) +svuint32_t svldff1_gather_s32index_u32(svbool_t, uint32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s32index_f32))) +svfloat32_t svldff1_gather_s32index_f32(svbool_t, float32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s32index_s32))) +svint32_t svldff1_gather_s32index_s32(svbool_t, int32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32index_u32))) +svuint32_t svldff1_gather_u32index_u32(svbool_t, uint32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32index_f32))) +svfloat32_t svldff1_gather_u32index_f32(svbool_t, float32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32index_s32))) +svint32_t svldff1_gather_u32index_s32(svbool_t, int32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s64index_u64))) +svuint64_t svldff1_gather_s64index_u64(svbool_t, uint64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s64index_f64))) +svfloat64_t svldff1_gather_s64index_f64(svbool_t, float64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s64index_s64))) +svint64_t svldff1_gather_s64index_s64(svbool_t, int64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64index_u64))) +svuint64_t svldff1_gather_u64index_u64(svbool_t, uint64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64index_f64))) +svfloat64_t svldff1_gather_u64index_f64(svbool_t, float64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64index_s64))) +svint64_t svldff1_gather_u64index_s64(svbool_t, int64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s32offset_u32))) +svuint32_t svldff1_gather_s32offset_u32(svbool_t, uint32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s32offset_f32))) +svfloat32_t svldff1_gather_s32offset_f32(svbool_t, float32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s32offset_s32))) +svint32_t svldff1_gather_s32offset_s32(svbool_t, int32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32offset_u32))) +svuint32_t svldff1_gather_u32offset_u32(svbool_t, uint32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32offset_f32))) +svfloat32_t svldff1_gather_u32offset_f32(svbool_t, float32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32offset_s32))) +svint32_t svldff1_gather_u32offset_s32(svbool_t, int32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s64offset_u64))) +svuint64_t svldff1_gather_s64offset_u64(svbool_t, uint64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s64offset_f64))) +svfloat64_t svldff1_gather_s64offset_f64(svbool_t, float64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s64offset_s64))) +svint64_t svldff1_gather_s64offset_s64(svbool_t, int64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64offset_u64))) +svuint64_t svldff1_gather_u64offset_u64(svbool_t, uint64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64offset_f64))) +svfloat64_t svldff1_gather_u64offset_f64(svbool_t, float64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64offset_s64))) +svint64_t svldff1_gather_u64offset_s64(svbool_t, int64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_u8))) +svuint8_t svldff1_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_u32))) +svuint32_t svldff1_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_u64))) +svuint64_t svldff1_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_u16))) +svuint16_t svldff1_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_s8))) +svint8_t svldff1_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_f64))) +svfloat64_t svldff1_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_f32))) +svfloat32_t svldff1_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_f16))) +svfloat16_t svldff1_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_s32))) +svint32_t svldff1_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_s64))) +svint64_t svldff1_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_s16))) +svint16_t svldff1_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u32base_offset_u32))) +svuint32_t svldff1sb_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u64base_offset_u64))) +svuint64_t svldff1sb_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u32base_offset_s32))) +svint32_t svldff1sb_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u64base_offset_s64))) +svint64_t svldff1sb_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u32base_u32))) +svuint32_t svldff1sb_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u64base_u64))) +svuint64_t svldff1sb_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u32base_s32))) +svint32_t svldff1sb_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u64base_s64))) +svint64_t svldff1sb_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_s32offset_u32))) +svuint32_t svldff1sb_gather_s32offset_u32(svbool_t, int8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_s32offset_s32))) +svint32_t svldff1sb_gather_s32offset_s32(svbool_t, int8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u32offset_u32))) +svuint32_t svldff1sb_gather_u32offset_u32(svbool_t, int8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u32offset_s32))) +svint32_t svldff1sb_gather_u32offset_s32(svbool_t, int8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_s64offset_u64))) +svuint64_t svldff1sb_gather_s64offset_u64(svbool_t, int8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_s64offset_s64))) +svint64_t svldff1sb_gather_s64offset_s64(svbool_t, int8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u64offset_u64))) +svuint64_t svldff1sb_gather_u64offset_u64(svbool_t, int8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u64offset_s64))) +svint64_t svldff1sb_gather_u64offset_s64(svbool_t, int8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_vnum_u32))) +svuint32_t svldff1sb_vnum_u32(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_vnum_u64))) +svuint64_t svldff1sb_vnum_u64(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_vnum_u16))) +svuint16_t svldff1sb_vnum_u16(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_vnum_s32))) +svint32_t svldff1sb_vnum_s32(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_vnum_s64))) +svint64_t svldff1sb_vnum_s64(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_vnum_s16))) +svint16_t svldff1sb_vnum_s16(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_u32))) +svuint32_t svldff1sb_u32(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_u64))) +svuint64_t svldff1sb_u64(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_u16))) +svuint16_t svldff1sb_u16(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_s32))) +svint32_t svldff1sb_s32(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_s64))) +svint64_t svldff1sb_s64(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_s16))) +svint16_t svldff1sb_s16(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32base_index_u32))) +svuint32_t svldff1sh_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64base_index_u64))) +svuint64_t svldff1sh_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32base_index_s32))) +svint32_t svldff1sh_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64base_index_s64))) +svint64_t svldff1sh_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32base_offset_u32))) +svuint32_t svldff1sh_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64base_offset_u64))) +svuint64_t svldff1sh_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32base_offset_s32))) +svint32_t svldff1sh_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64base_offset_s64))) +svint64_t svldff1sh_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32base_u32))) +svuint32_t svldff1sh_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64base_u64))) +svuint64_t svldff1sh_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32base_s32))) +svint32_t svldff1sh_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64base_s64))) +svint64_t svldff1sh_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s32index_u32))) +svuint32_t svldff1sh_gather_s32index_u32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s32index_s32))) +svint32_t svldff1sh_gather_s32index_s32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32index_u32))) +svuint32_t svldff1sh_gather_u32index_u32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32index_s32))) +svint32_t svldff1sh_gather_u32index_s32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s64index_u64))) +svuint64_t svldff1sh_gather_s64index_u64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s64index_s64))) +svint64_t svldff1sh_gather_s64index_s64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64index_u64))) +svuint64_t svldff1sh_gather_u64index_u64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64index_s64))) +svint64_t svldff1sh_gather_u64index_s64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s32offset_u32))) +svuint32_t svldff1sh_gather_s32offset_u32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s32offset_s32))) +svint32_t svldff1sh_gather_s32offset_s32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32offset_u32))) +svuint32_t svldff1sh_gather_u32offset_u32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32offset_s32))) +svint32_t svldff1sh_gather_u32offset_s32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s64offset_u64))) +svuint64_t svldff1sh_gather_s64offset_u64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s64offset_s64))) +svint64_t svldff1sh_gather_s64offset_s64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64offset_u64))) +svuint64_t svldff1sh_gather_u64offset_u64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64offset_s64))) +svint64_t svldff1sh_gather_u64offset_s64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_vnum_u32))) +svuint32_t svldff1sh_vnum_u32(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_vnum_u64))) +svuint64_t svldff1sh_vnum_u64(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_vnum_s32))) +svint32_t svldff1sh_vnum_s32(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_vnum_s64))) +svint64_t svldff1sh_vnum_s64(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_u32))) +svuint32_t svldff1sh_u32(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_u64))) +svuint64_t svldff1sh_u64(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_s32))) +svint32_t svldff1sh_s32(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_s64))) +svint64_t svldff1sh_s64(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64base_index_u64))) +svuint64_t svldff1sw_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64base_index_s64))) +svint64_t svldff1sw_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64base_offset_u64))) +svuint64_t svldff1sw_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64base_offset_s64))) +svint64_t svldff1sw_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64base_u64))) +svuint64_t svldff1sw_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64base_s64))) +svint64_t svldff1sw_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_s64index_u64))) +svuint64_t svldff1sw_gather_s64index_u64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_s64index_s64))) +svint64_t svldff1sw_gather_s64index_s64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64index_u64))) +svuint64_t svldff1sw_gather_u64index_u64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64index_s64))) +svint64_t svldff1sw_gather_u64index_s64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_s64offset_u64))) +svuint64_t svldff1sw_gather_s64offset_u64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_s64offset_s64))) +svint64_t svldff1sw_gather_s64offset_s64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64offset_u64))) +svuint64_t svldff1sw_gather_u64offset_u64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64offset_s64))) +svint64_t svldff1sw_gather_u64offset_s64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_vnum_u64))) +svuint64_t svldff1sw_vnum_u64(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_vnum_s64))) +svint64_t svldff1sw_vnum_s64(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_u64))) +svuint64_t svldff1sw_u64(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_s64))) +svint64_t svldff1sw_s64(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u32base_offset_u32))) +svuint32_t svldff1ub_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u64base_offset_u64))) +svuint64_t svldff1ub_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u32base_offset_s32))) +svint32_t svldff1ub_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u64base_offset_s64))) +svint64_t svldff1ub_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u32base_u32))) +svuint32_t svldff1ub_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u64base_u64))) +svuint64_t svldff1ub_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u32base_s32))) +svint32_t svldff1ub_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u64base_s64))) +svint64_t svldff1ub_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_s32offset_u32))) +svuint32_t svldff1ub_gather_s32offset_u32(svbool_t, uint8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_s32offset_s32))) +svint32_t svldff1ub_gather_s32offset_s32(svbool_t, uint8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u32offset_u32))) +svuint32_t svldff1ub_gather_u32offset_u32(svbool_t, uint8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u32offset_s32))) +svint32_t svldff1ub_gather_u32offset_s32(svbool_t, uint8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_s64offset_u64))) +svuint64_t svldff1ub_gather_s64offset_u64(svbool_t, uint8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_s64offset_s64))) +svint64_t svldff1ub_gather_s64offset_s64(svbool_t, uint8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u64offset_u64))) +svuint64_t svldff1ub_gather_u64offset_u64(svbool_t, uint8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u64offset_s64))) +svint64_t svldff1ub_gather_u64offset_s64(svbool_t, uint8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_vnum_u32))) +svuint32_t svldff1ub_vnum_u32(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_vnum_u64))) +svuint64_t svldff1ub_vnum_u64(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_vnum_u16))) +svuint16_t svldff1ub_vnum_u16(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_vnum_s32))) +svint32_t svldff1ub_vnum_s32(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_vnum_s64))) +svint64_t svldff1ub_vnum_s64(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_vnum_s16))) +svint16_t svldff1ub_vnum_s16(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_u32))) +svuint32_t svldff1ub_u32(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_u64))) +svuint64_t svldff1ub_u64(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_u16))) +svuint16_t svldff1ub_u16(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_s32))) +svint32_t svldff1ub_s32(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_s64))) +svint64_t svldff1ub_s64(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_s16))) +svint16_t svldff1ub_s16(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32base_index_u32))) +svuint32_t svldff1uh_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64base_index_u64))) +svuint64_t svldff1uh_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32base_index_s32))) +svint32_t svldff1uh_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64base_index_s64))) +svint64_t svldff1uh_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32base_offset_u32))) +svuint32_t svldff1uh_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64base_offset_u64))) +svuint64_t svldff1uh_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32base_offset_s32))) +svint32_t svldff1uh_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64base_offset_s64))) +svint64_t svldff1uh_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32base_u32))) +svuint32_t svldff1uh_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64base_u64))) +svuint64_t svldff1uh_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32base_s32))) +svint32_t svldff1uh_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64base_s64))) +svint64_t svldff1uh_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s32index_u32))) +svuint32_t svldff1uh_gather_s32index_u32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s32index_s32))) +svint32_t svldff1uh_gather_s32index_s32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32index_u32))) +svuint32_t svldff1uh_gather_u32index_u32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32index_s32))) +svint32_t svldff1uh_gather_u32index_s32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s64index_u64))) +svuint64_t svldff1uh_gather_s64index_u64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s64index_s64))) +svint64_t svldff1uh_gather_s64index_s64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64index_u64))) +svuint64_t svldff1uh_gather_u64index_u64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64index_s64))) +svint64_t svldff1uh_gather_u64index_s64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s32offset_u32))) +svuint32_t svldff1uh_gather_s32offset_u32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s32offset_s32))) +svint32_t svldff1uh_gather_s32offset_s32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32offset_u32))) +svuint32_t svldff1uh_gather_u32offset_u32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32offset_s32))) +svint32_t svldff1uh_gather_u32offset_s32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s64offset_u64))) +svuint64_t svldff1uh_gather_s64offset_u64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s64offset_s64))) +svint64_t svldff1uh_gather_s64offset_s64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64offset_u64))) +svuint64_t svldff1uh_gather_u64offset_u64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64offset_s64))) +svint64_t svldff1uh_gather_u64offset_s64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_vnum_u32))) +svuint32_t svldff1uh_vnum_u32(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_vnum_u64))) +svuint64_t svldff1uh_vnum_u64(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_vnum_s32))) +svint32_t svldff1uh_vnum_s32(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_vnum_s64))) +svint64_t svldff1uh_vnum_s64(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_u32))) +svuint32_t svldff1uh_u32(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_u64))) +svuint64_t svldff1uh_u64(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_s32))) +svint32_t svldff1uh_s32(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_s64))) +svint64_t svldff1uh_s64(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64base_index_u64))) +svuint64_t svldff1uw_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64base_index_s64))) +svint64_t svldff1uw_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64base_offset_u64))) +svuint64_t svldff1uw_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64base_offset_s64))) +svint64_t svldff1uw_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64base_u64))) +svuint64_t svldff1uw_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64base_s64))) +svint64_t svldff1uw_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_s64index_u64))) +svuint64_t svldff1uw_gather_s64index_u64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_s64index_s64))) +svint64_t svldff1uw_gather_s64index_s64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64index_u64))) +svuint64_t svldff1uw_gather_u64index_u64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64index_s64))) +svint64_t svldff1uw_gather_u64index_s64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_s64offset_u64))) +svuint64_t svldff1uw_gather_s64offset_u64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_s64offset_s64))) +svint64_t svldff1uw_gather_s64offset_s64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64offset_u64))) +svuint64_t svldff1uw_gather_u64offset_u64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64offset_s64))) +svint64_t svldff1uw_gather_u64offset_s64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_vnum_u64))) +svuint64_t svldff1uw_vnum_u64(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_vnum_s64))) +svint64_t svldff1uw_vnum_s64(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_u64))) +svuint64_t svldff1uw_u64(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_s64))) +svint64_t svldff1uw_s64(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_u8))) +svuint8_t svldnf1_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_u32))) +svuint32_t svldnf1_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_u64))) +svuint64_t svldnf1_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_u16))) +svuint16_t svldnf1_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_s8))) +svint8_t svldnf1_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_f64))) +svfloat64_t svldnf1_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_f32))) +svfloat32_t svldnf1_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_f16))) +svfloat16_t svldnf1_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_s32))) +svint32_t svldnf1_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_s64))) +svint64_t svldnf1_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_s16))) +svint16_t svldnf1_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_u8))) +svuint8_t svldnf1_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_u32))) +svuint32_t svldnf1_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_u64))) +svuint64_t svldnf1_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_u16))) +svuint16_t svldnf1_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_s8))) +svint8_t svldnf1_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_f64))) +svfloat64_t svldnf1_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_f32))) +svfloat32_t svldnf1_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_f16))) +svfloat16_t svldnf1_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_s32))) +svint32_t svldnf1_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_s64))) +svint64_t svldnf1_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_s16))) +svint16_t svldnf1_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_vnum_u32))) +svuint32_t svldnf1sb_vnum_u32(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_vnum_u64))) +svuint64_t svldnf1sb_vnum_u64(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_vnum_u16))) +svuint16_t svldnf1sb_vnum_u16(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_vnum_s32))) +svint32_t svldnf1sb_vnum_s32(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_vnum_s64))) +svint64_t svldnf1sb_vnum_s64(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_vnum_s16))) +svint16_t svldnf1sb_vnum_s16(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_u32))) +svuint32_t svldnf1sb_u32(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_u64))) +svuint64_t svldnf1sb_u64(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_u16))) +svuint16_t svldnf1sb_u16(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_s32))) +svint32_t svldnf1sb_s32(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_s64))) +svint64_t svldnf1sb_s64(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_s16))) +svint16_t svldnf1sb_s16(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_vnum_u32))) +svuint32_t svldnf1sh_vnum_u32(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_vnum_u64))) +svuint64_t svldnf1sh_vnum_u64(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_vnum_s32))) +svint32_t svldnf1sh_vnum_s32(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_vnum_s64))) +svint64_t svldnf1sh_vnum_s64(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_u32))) +svuint32_t svldnf1sh_u32(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_u64))) +svuint64_t svldnf1sh_u64(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_s32))) +svint32_t svldnf1sh_s32(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_s64))) +svint64_t svldnf1sh_s64(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sw_vnum_u64))) +svuint64_t svldnf1sw_vnum_u64(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sw_vnum_s64))) +svint64_t svldnf1sw_vnum_s64(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sw_u64))) +svuint64_t svldnf1sw_u64(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sw_s64))) +svint64_t svldnf1sw_s64(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_vnum_u32))) +svuint32_t svldnf1ub_vnum_u32(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_vnum_u64))) +svuint64_t svldnf1ub_vnum_u64(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_vnum_u16))) +svuint16_t svldnf1ub_vnum_u16(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_vnum_s32))) +svint32_t svldnf1ub_vnum_s32(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_vnum_s64))) +svint64_t svldnf1ub_vnum_s64(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_vnum_s16))) +svint16_t svldnf1ub_vnum_s16(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_u32))) +svuint32_t svldnf1ub_u32(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_u64))) +svuint64_t svldnf1ub_u64(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_u16))) +svuint16_t svldnf1ub_u16(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_s32))) +svint32_t svldnf1ub_s32(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_s64))) +svint64_t svldnf1ub_s64(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_s16))) +svint16_t svldnf1ub_s16(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_vnum_u32))) +svuint32_t svldnf1uh_vnum_u32(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_vnum_u64))) +svuint64_t svldnf1uh_vnum_u64(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_vnum_s32))) +svint32_t svldnf1uh_vnum_s32(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_vnum_s64))) +svint64_t svldnf1uh_vnum_s64(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_u32))) +svuint32_t svldnf1uh_u32(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_u64))) +svuint64_t svldnf1uh_u64(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_s32))) +svint32_t svldnf1uh_s32(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_s64))) +svint64_t svldnf1uh_s64(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uw_vnum_u64))) +svuint64_t svldnf1uw_vnum_u64(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uw_vnum_s64))) +svint64_t svldnf1uw_vnum_s64(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uw_u64))) +svuint64_t svldnf1uw_u64(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uw_s64))) +svint64_t svldnf1uw_s64(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_u8))) +svuint8_t svldnt1_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_u32))) +svuint32_t svldnt1_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_u64))) +svuint64_t svldnt1_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_u16))) +svuint16_t svldnt1_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_s8))) +svint8_t svldnt1_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_f64))) +svfloat64_t svldnt1_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_f32))) +svfloat32_t svldnt1_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_f16))) +svfloat16_t svldnt1_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_s32))) +svint32_t svldnt1_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_s64))) +svint64_t svldnt1_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_s16))) +svint16_t svldnt1_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_u8))) +svuint8_t svldnt1_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_u32))) +svuint32_t svldnt1_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_u64))) +svuint64_t svldnt1_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_u16))) +svuint16_t svldnt1_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_s8))) +svint8_t svldnt1_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_f64))) +svfloat64_t svldnt1_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_f32))) +svfloat32_t svldnt1_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_f16))) +svfloat16_t svldnt1_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_s32))) +svint32_t svldnt1_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_s64))) +svint64_t svldnt1_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_s16))) +svint16_t svldnt1_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_u8))) +uint64_t svlen_u8(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_u32))) +uint64_t svlen_u32(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_u64))) +uint64_t svlen_u64(svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_u16))) +uint64_t svlen_u16(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_s8))) +uint64_t svlen_s8(svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_f64))) +uint64_t svlen_f64(svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_f32))) +uint64_t svlen_f32(svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_f16))) +uint64_t svlen_f16(svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_s32))) +uint64_t svlen_s32(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_s64))) +uint64_t svlen_s64(svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_s16))) +uint64_t svlen_s16(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u8_m))) +svuint8_t svlsl_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u32_m))) +svuint32_t svlsl_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u64_m))) +svuint64_t svlsl_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u16_m))) +svuint16_t svlsl_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s8_m))) +svint8_t svlsl_n_s8_m(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s32_m))) +svint32_t svlsl_n_s32_m(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s64_m))) +svint64_t svlsl_n_s64_m(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s16_m))) +svint16_t svlsl_n_s16_m(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u8_x))) +svuint8_t svlsl_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u32_x))) +svuint32_t svlsl_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u64_x))) +svuint64_t svlsl_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u16_x))) +svuint16_t svlsl_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s8_x))) +svint8_t svlsl_n_s8_x(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s32_x))) +svint32_t svlsl_n_s32_x(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s64_x))) +svint64_t svlsl_n_s64_x(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s16_x))) +svint16_t svlsl_n_s16_x(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u8_z))) +svuint8_t svlsl_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u32_z))) +svuint32_t svlsl_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u64_z))) +svuint64_t svlsl_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u16_z))) +svuint16_t svlsl_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s8_z))) +svint8_t svlsl_n_s8_z(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s32_z))) +svint32_t svlsl_n_s32_z(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s64_z))) +svint64_t svlsl_n_s64_z(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s16_z))) +svint16_t svlsl_n_s16_z(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u8_m))) +svuint8_t svlsl_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u32_m))) +svuint32_t svlsl_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u64_m))) +svuint64_t svlsl_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u16_m))) +svuint16_t svlsl_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s8_m))) +svint8_t svlsl_s8_m(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s32_m))) +svint32_t svlsl_s32_m(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s64_m))) +svint64_t svlsl_s64_m(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s16_m))) +svint16_t svlsl_s16_m(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u8_x))) +svuint8_t svlsl_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u32_x))) +svuint32_t svlsl_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u64_x))) +svuint64_t svlsl_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u16_x))) +svuint16_t svlsl_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s8_x))) +svint8_t svlsl_s8_x(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s32_x))) +svint32_t svlsl_s32_x(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s64_x))) +svint64_t svlsl_s64_x(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s16_x))) +svint16_t svlsl_s16_x(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u8_z))) +svuint8_t svlsl_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u32_z))) +svuint32_t svlsl_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u64_z))) +svuint64_t svlsl_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u16_z))) +svuint16_t svlsl_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s8_z))) +svint8_t svlsl_s8_z(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s32_z))) +svint32_t svlsl_s32_z(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s64_z))) +svint64_t svlsl_s64_z(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s16_z))) +svint16_t svlsl_s16_z(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u8_m))) +svuint8_t svlsl_wide_n_u8_m(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u32_m))) +svuint32_t svlsl_wide_n_u32_m(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u16_m))) +svuint16_t svlsl_wide_n_u16_m(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s8_m))) +svint8_t svlsl_wide_n_s8_m(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s32_m))) +svint32_t svlsl_wide_n_s32_m(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s16_m))) +svint16_t svlsl_wide_n_s16_m(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u8_x))) +svuint8_t svlsl_wide_n_u8_x(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u32_x))) +svuint32_t svlsl_wide_n_u32_x(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u16_x))) +svuint16_t svlsl_wide_n_u16_x(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s8_x))) +svint8_t svlsl_wide_n_s8_x(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s32_x))) +svint32_t svlsl_wide_n_s32_x(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s16_x))) +svint16_t svlsl_wide_n_s16_x(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u8_z))) +svuint8_t svlsl_wide_n_u8_z(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u32_z))) +svuint32_t svlsl_wide_n_u32_z(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u16_z))) +svuint16_t svlsl_wide_n_u16_z(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s8_z))) +svint8_t svlsl_wide_n_s8_z(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s32_z))) +svint32_t svlsl_wide_n_s32_z(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s16_z))) +svint16_t svlsl_wide_n_s16_z(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u8_m))) +svuint8_t svlsl_wide_u8_m(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u32_m))) +svuint32_t svlsl_wide_u32_m(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u16_m))) +svuint16_t svlsl_wide_u16_m(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s8_m))) +svint8_t svlsl_wide_s8_m(svbool_t, svint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s32_m))) +svint32_t svlsl_wide_s32_m(svbool_t, svint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s16_m))) +svint16_t svlsl_wide_s16_m(svbool_t, svint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u8_x))) +svuint8_t svlsl_wide_u8_x(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u32_x))) +svuint32_t svlsl_wide_u32_x(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u16_x))) +svuint16_t svlsl_wide_u16_x(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s8_x))) +svint8_t svlsl_wide_s8_x(svbool_t, svint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s32_x))) +svint32_t svlsl_wide_s32_x(svbool_t, svint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s16_x))) +svint16_t svlsl_wide_s16_x(svbool_t, svint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u8_z))) +svuint8_t svlsl_wide_u8_z(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u32_z))) +svuint32_t svlsl_wide_u32_z(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u16_z))) +svuint16_t svlsl_wide_u16_z(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s8_z))) +svint8_t svlsl_wide_s8_z(svbool_t, svint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s32_z))) +svint32_t svlsl_wide_s32_z(svbool_t, svint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s16_z))) +svint16_t svlsl_wide_s16_z(svbool_t, svint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u8_m))) +svuint8_t svlsr_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u32_m))) +svuint32_t svlsr_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u64_m))) +svuint64_t svlsr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u16_m))) +svuint16_t svlsr_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u8_x))) +svuint8_t svlsr_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u32_x))) +svuint32_t svlsr_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u64_x))) +svuint64_t svlsr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u16_x))) +svuint16_t svlsr_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u8_z))) +svuint8_t svlsr_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u32_z))) +svuint32_t svlsr_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u64_z))) +svuint64_t svlsr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u16_z))) +svuint16_t svlsr_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u8_m))) +svuint8_t svlsr_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u32_m))) +svuint32_t svlsr_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u64_m))) +svuint64_t svlsr_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u16_m))) +svuint16_t svlsr_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u8_x))) +svuint8_t svlsr_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u32_x))) +svuint32_t svlsr_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u64_x))) +svuint64_t svlsr_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u16_x))) +svuint16_t svlsr_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u8_z))) +svuint8_t svlsr_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u32_z))) +svuint32_t svlsr_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u64_z))) +svuint64_t svlsr_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u16_z))) +svuint16_t svlsr_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u8_m))) +svuint8_t svlsr_wide_n_u8_m(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u32_m))) +svuint32_t svlsr_wide_n_u32_m(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u16_m))) +svuint16_t svlsr_wide_n_u16_m(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u8_x))) +svuint8_t svlsr_wide_n_u8_x(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u32_x))) +svuint32_t svlsr_wide_n_u32_x(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u16_x))) +svuint16_t svlsr_wide_n_u16_x(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u8_z))) +svuint8_t svlsr_wide_n_u8_z(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u32_z))) +svuint32_t svlsr_wide_n_u32_z(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u16_z))) +svuint16_t svlsr_wide_n_u16_z(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u8_m))) +svuint8_t svlsr_wide_u8_m(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u32_m))) +svuint32_t svlsr_wide_u32_m(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u16_m))) +svuint16_t svlsr_wide_u16_m(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u8_x))) +svuint8_t svlsr_wide_u8_x(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u32_x))) +svuint32_t svlsr_wide_u32_x(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u16_x))) +svuint16_t svlsr_wide_u16_x(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u8_z))) +svuint8_t svlsr_wide_u8_z(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u32_z))) +svuint32_t svlsr_wide_u32_z(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u16_z))) +svuint16_t svlsr_wide_u16_z(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f64_m))) +svfloat64_t svmad_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f32_m))) +svfloat32_t svmad_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f16_m))) +svfloat16_t svmad_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f64_x))) +svfloat64_t svmad_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f32_x))) +svfloat32_t svmad_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f16_x))) +svfloat16_t svmad_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f64_z))) +svfloat64_t svmad_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f32_z))) +svfloat32_t svmad_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f16_z))) +svfloat16_t svmad_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u8_m))) +svuint8_t svmad_n_u8_m(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u32_m))) +svuint32_t svmad_n_u32_m(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u64_m))) +svuint64_t svmad_n_u64_m(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u16_m))) +svuint16_t svmad_n_u16_m(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s8_m))) +svint8_t svmad_n_s8_m(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s32_m))) +svint32_t svmad_n_s32_m(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s64_m))) +svint64_t svmad_n_s64_m(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s16_m))) +svint16_t svmad_n_s16_m(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u8_x))) +svuint8_t svmad_n_u8_x(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u32_x))) +svuint32_t svmad_n_u32_x(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u64_x))) +svuint64_t svmad_n_u64_x(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u16_x))) +svuint16_t svmad_n_u16_x(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s8_x))) +svint8_t svmad_n_s8_x(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s32_x))) +svint32_t svmad_n_s32_x(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s64_x))) +svint64_t svmad_n_s64_x(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s16_x))) +svint16_t svmad_n_s16_x(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u8_z))) +svuint8_t svmad_n_u8_z(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u32_z))) +svuint32_t svmad_n_u32_z(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u64_z))) +svuint64_t svmad_n_u64_z(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u16_z))) +svuint16_t svmad_n_u16_z(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s8_z))) +svint8_t svmad_n_s8_z(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s32_z))) +svint32_t svmad_n_s32_z(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s64_z))) +svint64_t svmad_n_s64_z(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s16_z))) +svint16_t svmad_n_s16_z(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f64_m))) +svfloat64_t svmad_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f32_m))) +svfloat32_t svmad_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f16_m))) +svfloat16_t svmad_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f64_x))) +svfloat64_t svmad_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f32_x))) +svfloat32_t svmad_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f16_x))) +svfloat16_t svmad_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f64_z))) +svfloat64_t svmad_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f32_z))) +svfloat32_t svmad_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f16_z))) +svfloat16_t svmad_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u8_m))) +svuint8_t svmad_u8_m(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u32_m))) +svuint32_t svmad_u32_m(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u64_m))) +svuint64_t svmad_u64_m(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u16_m))) +svuint16_t svmad_u16_m(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s8_m))) +svint8_t svmad_s8_m(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s32_m))) +svint32_t svmad_s32_m(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s64_m))) +svint64_t svmad_s64_m(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s16_m))) +svint16_t svmad_s16_m(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u8_x))) +svuint8_t svmad_u8_x(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u32_x))) +svuint32_t svmad_u32_x(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u64_x))) +svuint64_t svmad_u64_x(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u16_x))) +svuint16_t svmad_u16_x(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s8_x))) +svint8_t svmad_s8_x(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s32_x))) +svint32_t svmad_s32_x(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s64_x))) +svint64_t svmad_s64_x(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s16_x))) +svint16_t svmad_s16_x(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u8_z))) +svuint8_t svmad_u8_z(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u32_z))) +svuint32_t svmad_u32_z(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u64_z))) +svuint64_t svmad_u64_z(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u16_z))) +svuint16_t svmad_u16_z(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s8_z))) +svint8_t svmad_s8_z(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s32_z))) +svint32_t svmad_s32_z(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s64_z))) +svint64_t svmad_s64_z(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s16_z))) +svint16_t svmad_s16_z(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f64_m))) +svfloat64_t svmax_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f32_m))) +svfloat32_t svmax_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f16_m))) +svfloat16_t svmax_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f64_x))) +svfloat64_t svmax_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f32_x))) +svfloat32_t svmax_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f16_x))) +svfloat16_t svmax_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f64_z))) +svfloat64_t svmax_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f32_z))) +svfloat32_t svmax_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f16_z))) +svfloat16_t svmax_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s8_m))) +svint8_t svmax_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s32_m))) +svint32_t svmax_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s64_m))) +svint64_t svmax_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s16_m))) +svint16_t svmax_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s8_x))) +svint8_t svmax_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s32_x))) +svint32_t svmax_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s64_x))) +svint64_t svmax_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s16_x))) +svint16_t svmax_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s8_z))) +svint8_t svmax_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s32_z))) +svint32_t svmax_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s64_z))) +svint64_t svmax_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s16_z))) +svint16_t svmax_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u8_m))) +svuint8_t svmax_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u32_m))) +svuint32_t svmax_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u64_m))) +svuint64_t svmax_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u16_m))) +svuint16_t svmax_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u8_x))) +svuint8_t svmax_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u32_x))) +svuint32_t svmax_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u64_x))) +svuint64_t svmax_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u16_x))) +svuint16_t svmax_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u8_z))) +svuint8_t svmax_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u32_z))) +svuint32_t svmax_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u64_z))) +svuint64_t svmax_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u16_z))) +svuint16_t svmax_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f64_m))) +svfloat64_t svmax_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f32_m))) +svfloat32_t svmax_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f16_m))) +svfloat16_t svmax_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f64_x))) +svfloat64_t svmax_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f32_x))) +svfloat32_t svmax_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f16_x))) +svfloat16_t svmax_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f64_z))) +svfloat64_t svmax_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f32_z))) +svfloat32_t svmax_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f16_z))) +svfloat16_t svmax_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s8_m))) +svint8_t svmax_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s32_m))) +svint32_t svmax_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s64_m))) +svint64_t svmax_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s16_m))) +svint16_t svmax_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s8_x))) +svint8_t svmax_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s32_x))) +svint32_t svmax_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s64_x))) +svint64_t svmax_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s16_x))) +svint16_t svmax_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s8_z))) +svint8_t svmax_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s32_z))) +svint32_t svmax_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s64_z))) +svint64_t svmax_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s16_z))) +svint16_t svmax_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u8_m))) +svuint8_t svmax_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u32_m))) +svuint32_t svmax_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u64_m))) +svuint64_t svmax_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u16_m))) +svuint16_t svmax_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u8_x))) +svuint8_t svmax_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u32_x))) +svuint32_t svmax_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u64_x))) +svuint64_t svmax_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u16_x))) +svuint16_t svmax_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u8_z))) +svuint8_t svmax_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u32_z))) +svuint32_t svmax_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u64_z))) +svuint64_t svmax_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u16_z))) +svuint16_t svmax_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f64_m))) +svfloat64_t svmaxnm_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f32_m))) +svfloat32_t svmaxnm_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f16_m))) +svfloat16_t svmaxnm_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f64_x))) +svfloat64_t svmaxnm_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f32_x))) +svfloat32_t svmaxnm_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f16_x))) +svfloat16_t svmaxnm_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f64_z))) +svfloat64_t svmaxnm_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f32_z))) +svfloat32_t svmaxnm_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f16_z))) +svfloat16_t svmaxnm_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f64_m))) +svfloat64_t svmaxnm_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f32_m))) +svfloat32_t svmaxnm_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f16_m))) +svfloat16_t svmaxnm_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f64_x))) +svfloat64_t svmaxnm_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f32_x))) +svfloat32_t svmaxnm_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f16_x))) +svfloat16_t svmaxnm_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f64_z))) +svfloat64_t svmaxnm_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f32_z))) +svfloat32_t svmaxnm_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f16_z))) +svfloat16_t svmaxnm_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmv_f64))) +float64_t svmaxnmv_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmv_f32))) +float32_t svmaxnmv_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmv_f16))) +float16_t svmaxnmv_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_f64))) +float64_t svmaxv_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_f32))) +float32_t svmaxv_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_f16))) +float16_t svmaxv_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_s8))) +int8_t svmaxv_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_s32))) +int32_t svmaxv_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_s64))) +int64_t svmaxv_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_s16))) +int16_t svmaxv_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_u8))) +uint8_t svmaxv_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_u32))) +uint32_t svmaxv_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_u64))) +uint64_t svmaxv_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_u16))) +uint16_t svmaxv_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f64_m))) +svfloat64_t svmin_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f32_m))) +svfloat32_t svmin_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f16_m))) +svfloat16_t svmin_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f64_x))) +svfloat64_t svmin_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f32_x))) +svfloat32_t svmin_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f16_x))) +svfloat16_t svmin_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f64_z))) +svfloat64_t svmin_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f32_z))) +svfloat32_t svmin_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f16_z))) +svfloat16_t svmin_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s8_m))) +svint8_t svmin_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s32_m))) +svint32_t svmin_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s64_m))) +svint64_t svmin_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s16_m))) +svint16_t svmin_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s8_x))) +svint8_t svmin_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s32_x))) +svint32_t svmin_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s64_x))) +svint64_t svmin_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s16_x))) +svint16_t svmin_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s8_z))) +svint8_t svmin_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s32_z))) +svint32_t svmin_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s64_z))) +svint64_t svmin_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s16_z))) +svint16_t svmin_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u8_m))) +svuint8_t svmin_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u32_m))) +svuint32_t svmin_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u64_m))) +svuint64_t svmin_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u16_m))) +svuint16_t svmin_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u8_x))) +svuint8_t svmin_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u32_x))) +svuint32_t svmin_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u64_x))) +svuint64_t svmin_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u16_x))) +svuint16_t svmin_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u8_z))) +svuint8_t svmin_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u32_z))) +svuint32_t svmin_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u64_z))) +svuint64_t svmin_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u16_z))) +svuint16_t svmin_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f64_m))) +svfloat64_t svmin_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f32_m))) +svfloat32_t svmin_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f16_m))) +svfloat16_t svmin_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f64_x))) +svfloat64_t svmin_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f32_x))) +svfloat32_t svmin_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f16_x))) +svfloat16_t svmin_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f64_z))) +svfloat64_t svmin_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f32_z))) +svfloat32_t svmin_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f16_z))) +svfloat16_t svmin_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s8_m))) +svint8_t svmin_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s32_m))) +svint32_t svmin_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s64_m))) +svint64_t svmin_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s16_m))) +svint16_t svmin_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s8_x))) +svint8_t svmin_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s32_x))) +svint32_t svmin_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s64_x))) +svint64_t svmin_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s16_x))) +svint16_t svmin_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s8_z))) +svint8_t svmin_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s32_z))) +svint32_t svmin_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s64_z))) +svint64_t svmin_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s16_z))) +svint16_t svmin_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u8_m))) +svuint8_t svmin_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u32_m))) +svuint32_t svmin_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u64_m))) +svuint64_t svmin_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u16_m))) +svuint16_t svmin_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u8_x))) +svuint8_t svmin_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u32_x))) +svuint32_t svmin_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u64_x))) +svuint64_t svmin_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u16_x))) +svuint16_t svmin_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u8_z))) +svuint8_t svmin_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u32_z))) +svuint32_t svmin_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u64_z))) +svuint64_t svmin_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u16_z))) +svuint16_t svmin_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f64_m))) +svfloat64_t svminnm_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f32_m))) +svfloat32_t svminnm_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f16_m))) +svfloat16_t svminnm_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f64_x))) +svfloat64_t svminnm_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f32_x))) +svfloat32_t svminnm_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f16_x))) +svfloat16_t svminnm_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f64_z))) +svfloat64_t svminnm_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f32_z))) +svfloat32_t svminnm_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f16_z))) +svfloat16_t svminnm_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f64_m))) +svfloat64_t svminnm_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f32_m))) +svfloat32_t svminnm_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f16_m))) +svfloat16_t svminnm_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f64_x))) +svfloat64_t svminnm_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f32_x))) +svfloat32_t svminnm_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f16_x))) +svfloat16_t svminnm_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f64_z))) +svfloat64_t svminnm_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f32_z))) +svfloat32_t svminnm_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f16_z))) +svfloat16_t svminnm_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmv_f64))) +float64_t svminnmv_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmv_f32))) +float32_t svminnmv_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmv_f16))) +float16_t svminnmv_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_f64))) +float64_t svminv_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_f32))) +float32_t svminv_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_f16))) +float16_t svminv_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_s8))) +int8_t svminv_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_s32))) +int32_t svminv_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_s64))) +int64_t svminv_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_s16))) +int16_t svminv_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_u8))) +uint8_t svminv_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_u32))) +uint32_t svminv_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_u64))) +uint64_t svminv_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_u16))) +uint16_t svminv_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f64_m))) +svfloat64_t svmla_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f32_m))) +svfloat32_t svmla_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f16_m))) +svfloat16_t svmla_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f64_x))) +svfloat64_t svmla_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f32_x))) +svfloat32_t svmla_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f16_x))) +svfloat16_t svmla_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f64_z))) +svfloat64_t svmla_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f32_z))) +svfloat32_t svmla_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f16_z))) +svfloat16_t svmla_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u8_m))) +svuint8_t svmla_n_u8_m(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u32_m))) +svuint32_t svmla_n_u32_m(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u64_m))) +svuint64_t svmla_n_u64_m(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u16_m))) +svuint16_t svmla_n_u16_m(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s8_m))) +svint8_t svmla_n_s8_m(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s32_m))) +svint32_t svmla_n_s32_m(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s64_m))) +svint64_t svmla_n_s64_m(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s16_m))) +svint16_t svmla_n_s16_m(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u8_x))) +svuint8_t svmla_n_u8_x(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u32_x))) +svuint32_t svmla_n_u32_x(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u64_x))) +svuint64_t svmla_n_u64_x(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u16_x))) +svuint16_t svmla_n_u16_x(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s8_x))) +svint8_t svmla_n_s8_x(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s32_x))) +svint32_t svmla_n_s32_x(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s64_x))) +svint64_t svmla_n_s64_x(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s16_x))) +svint16_t svmla_n_s16_x(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u8_z))) +svuint8_t svmla_n_u8_z(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u32_z))) +svuint32_t svmla_n_u32_z(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u64_z))) +svuint64_t svmla_n_u64_z(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u16_z))) +svuint16_t svmla_n_u16_z(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s8_z))) +svint8_t svmla_n_s8_z(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s32_z))) +svint32_t svmla_n_s32_z(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s64_z))) +svint64_t svmla_n_s64_z(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s16_z))) +svint16_t svmla_n_s16_z(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f64_m))) +svfloat64_t svmla_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f32_m))) +svfloat32_t svmla_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f16_m))) +svfloat16_t svmla_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f64_x))) +svfloat64_t svmla_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f32_x))) +svfloat32_t svmla_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f16_x))) +svfloat16_t svmla_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f64_z))) +svfloat64_t svmla_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f32_z))) +svfloat32_t svmla_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f16_z))) +svfloat16_t svmla_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u8_m))) +svuint8_t svmla_u8_m(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u32_m))) +svuint32_t svmla_u32_m(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u64_m))) +svuint64_t svmla_u64_m(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u16_m))) +svuint16_t svmla_u16_m(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s8_m))) +svint8_t svmla_s8_m(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s32_m))) +svint32_t svmla_s32_m(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s64_m))) +svint64_t svmla_s64_m(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s16_m))) +svint16_t svmla_s16_m(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u8_x))) +svuint8_t svmla_u8_x(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u32_x))) +svuint32_t svmla_u32_x(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u64_x))) +svuint64_t svmla_u64_x(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u16_x))) +svuint16_t svmla_u16_x(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s8_x))) +svint8_t svmla_s8_x(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s32_x))) +svint32_t svmla_s32_x(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s64_x))) +svint64_t svmla_s64_x(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s16_x))) +svint16_t svmla_s16_x(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u8_z))) +svuint8_t svmla_u8_z(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u32_z))) +svuint32_t svmla_u32_z(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u64_z))) +svuint64_t svmla_u64_z(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u16_z))) +svuint16_t svmla_u16_z(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s8_z))) +svint8_t svmla_s8_z(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s32_z))) +svint32_t svmla_s32_z(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s64_z))) +svint64_t svmla_s64_z(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s16_z))) +svint16_t svmla_s16_z(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_f64))) +svfloat64_t svmla_lane_f64(svfloat64_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_f32))) +svfloat32_t svmla_lane_f32(svfloat32_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_f16))) +svfloat16_t svmla_lane_f16(svfloat16_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f64_m))) +svfloat64_t svmls_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f32_m))) +svfloat32_t svmls_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f16_m))) +svfloat16_t svmls_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f64_x))) +svfloat64_t svmls_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f32_x))) +svfloat32_t svmls_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f16_x))) +svfloat16_t svmls_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f64_z))) +svfloat64_t svmls_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f32_z))) +svfloat32_t svmls_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f16_z))) +svfloat16_t svmls_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u8_m))) +svuint8_t svmls_n_u8_m(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u32_m))) +svuint32_t svmls_n_u32_m(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u64_m))) +svuint64_t svmls_n_u64_m(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u16_m))) +svuint16_t svmls_n_u16_m(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s8_m))) +svint8_t svmls_n_s8_m(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s32_m))) +svint32_t svmls_n_s32_m(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s64_m))) +svint64_t svmls_n_s64_m(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s16_m))) +svint16_t svmls_n_s16_m(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u8_x))) +svuint8_t svmls_n_u8_x(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u32_x))) +svuint32_t svmls_n_u32_x(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u64_x))) +svuint64_t svmls_n_u64_x(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u16_x))) +svuint16_t svmls_n_u16_x(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s8_x))) +svint8_t svmls_n_s8_x(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s32_x))) +svint32_t svmls_n_s32_x(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s64_x))) +svint64_t svmls_n_s64_x(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s16_x))) +svint16_t svmls_n_s16_x(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u8_z))) +svuint8_t svmls_n_u8_z(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u32_z))) +svuint32_t svmls_n_u32_z(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u64_z))) +svuint64_t svmls_n_u64_z(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u16_z))) +svuint16_t svmls_n_u16_z(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s8_z))) +svint8_t svmls_n_s8_z(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s32_z))) +svint32_t svmls_n_s32_z(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s64_z))) +svint64_t svmls_n_s64_z(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s16_z))) +svint16_t svmls_n_s16_z(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f64_m))) +svfloat64_t svmls_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f32_m))) +svfloat32_t svmls_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f16_m))) +svfloat16_t svmls_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f64_x))) +svfloat64_t svmls_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f32_x))) +svfloat32_t svmls_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f16_x))) +svfloat16_t svmls_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f64_z))) +svfloat64_t svmls_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f32_z))) +svfloat32_t svmls_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f16_z))) +svfloat16_t svmls_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u8_m))) +svuint8_t svmls_u8_m(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u32_m))) +svuint32_t svmls_u32_m(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u64_m))) +svuint64_t svmls_u64_m(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u16_m))) +svuint16_t svmls_u16_m(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s8_m))) +svint8_t svmls_s8_m(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s32_m))) +svint32_t svmls_s32_m(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s64_m))) +svint64_t svmls_s64_m(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s16_m))) +svint16_t svmls_s16_m(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u8_x))) +svuint8_t svmls_u8_x(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u32_x))) +svuint32_t svmls_u32_x(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u64_x))) +svuint64_t svmls_u64_x(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u16_x))) +svuint16_t svmls_u16_x(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s8_x))) +svint8_t svmls_s8_x(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s32_x))) +svint32_t svmls_s32_x(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s64_x))) +svint64_t svmls_s64_x(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s16_x))) +svint16_t svmls_s16_x(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u8_z))) +svuint8_t svmls_u8_z(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u32_z))) +svuint32_t svmls_u32_z(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u64_z))) +svuint64_t svmls_u64_z(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u16_z))) +svuint16_t svmls_u16_z(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s8_z))) +svint8_t svmls_s8_z(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s32_z))) +svint32_t svmls_s32_z(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s64_z))) +svint64_t svmls_s64_z(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s16_z))) +svint16_t svmls_s16_z(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_f64))) +svfloat64_t svmls_lane_f64(svfloat64_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_f32))) +svfloat32_t svmls_lane_f32(svfloat32_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_f16))) +svfloat16_t svmls_lane_f16(svfloat16_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmov_b_z))) +svbool_t svmov_b_z(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f64_m))) +svfloat64_t svmsb_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f32_m))) +svfloat32_t svmsb_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f16_m))) +svfloat16_t svmsb_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f64_x))) +svfloat64_t svmsb_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f32_x))) +svfloat32_t svmsb_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f16_x))) +svfloat16_t svmsb_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f64_z))) +svfloat64_t svmsb_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f32_z))) +svfloat32_t svmsb_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f16_z))) +svfloat16_t svmsb_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u8_m))) +svuint8_t svmsb_n_u8_m(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u32_m))) +svuint32_t svmsb_n_u32_m(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u64_m))) +svuint64_t svmsb_n_u64_m(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u16_m))) +svuint16_t svmsb_n_u16_m(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s8_m))) +svint8_t svmsb_n_s8_m(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s32_m))) +svint32_t svmsb_n_s32_m(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s64_m))) +svint64_t svmsb_n_s64_m(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s16_m))) +svint16_t svmsb_n_s16_m(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u8_x))) +svuint8_t svmsb_n_u8_x(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u32_x))) +svuint32_t svmsb_n_u32_x(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u64_x))) +svuint64_t svmsb_n_u64_x(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u16_x))) +svuint16_t svmsb_n_u16_x(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s8_x))) +svint8_t svmsb_n_s8_x(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s32_x))) +svint32_t svmsb_n_s32_x(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s64_x))) +svint64_t svmsb_n_s64_x(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s16_x))) +svint16_t svmsb_n_s16_x(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u8_z))) +svuint8_t svmsb_n_u8_z(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u32_z))) +svuint32_t svmsb_n_u32_z(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u64_z))) +svuint64_t svmsb_n_u64_z(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u16_z))) +svuint16_t svmsb_n_u16_z(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s8_z))) +svint8_t svmsb_n_s8_z(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s32_z))) +svint32_t svmsb_n_s32_z(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s64_z))) +svint64_t svmsb_n_s64_z(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s16_z))) +svint16_t svmsb_n_s16_z(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f64_m))) +svfloat64_t svmsb_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f32_m))) +svfloat32_t svmsb_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f16_m))) +svfloat16_t svmsb_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f64_x))) +svfloat64_t svmsb_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f32_x))) +svfloat32_t svmsb_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f16_x))) +svfloat16_t svmsb_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f64_z))) +svfloat64_t svmsb_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f32_z))) +svfloat32_t svmsb_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f16_z))) +svfloat16_t svmsb_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u8_m))) +svuint8_t svmsb_u8_m(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u32_m))) +svuint32_t svmsb_u32_m(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u64_m))) +svuint64_t svmsb_u64_m(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u16_m))) +svuint16_t svmsb_u16_m(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s8_m))) +svint8_t svmsb_s8_m(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s32_m))) +svint32_t svmsb_s32_m(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s64_m))) +svint64_t svmsb_s64_m(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s16_m))) +svint16_t svmsb_s16_m(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u8_x))) +svuint8_t svmsb_u8_x(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u32_x))) +svuint32_t svmsb_u32_x(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u64_x))) +svuint64_t svmsb_u64_x(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u16_x))) +svuint16_t svmsb_u16_x(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s8_x))) +svint8_t svmsb_s8_x(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s32_x))) +svint32_t svmsb_s32_x(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s64_x))) +svint64_t svmsb_s64_x(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s16_x))) +svint16_t svmsb_s16_x(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u8_z))) +svuint8_t svmsb_u8_z(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u32_z))) +svuint32_t svmsb_u32_z(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u64_z))) +svuint64_t svmsb_u64_z(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u16_z))) +svuint16_t svmsb_u16_z(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s8_z))) +svint8_t svmsb_s8_z(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s32_z))) +svint32_t svmsb_s32_z(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s64_z))) +svint64_t svmsb_s64_z(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s16_z))) +svint16_t svmsb_s16_z(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f64_m))) +svfloat64_t svmul_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f32_m))) +svfloat32_t svmul_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f16_m))) +svfloat16_t svmul_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f64_x))) +svfloat64_t svmul_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f32_x))) +svfloat32_t svmul_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f16_x))) +svfloat16_t svmul_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f64_z))) +svfloat64_t svmul_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f32_z))) +svfloat32_t svmul_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f16_z))) +svfloat16_t svmul_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u8_m))) +svuint8_t svmul_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u32_m))) +svuint32_t svmul_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u64_m))) +svuint64_t svmul_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u16_m))) +svuint16_t svmul_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s8_m))) +svint8_t svmul_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s32_m))) +svint32_t svmul_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s64_m))) +svint64_t svmul_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s16_m))) +svint16_t svmul_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u8_x))) +svuint8_t svmul_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u32_x))) +svuint32_t svmul_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u64_x))) +svuint64_t svmul_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u16_x))) +svuint16_t svmul_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s8_x))) +svint8_t svmul_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s32_x))) +svint32_t svmul_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s64_x))) +svint64_t svmul_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s16_x))) +svint16_t svmul_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u8_z))) +svuint8_t svmul_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u32_z))) +svuint32_t svmul_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u64_z))) +svuint64_t svmul_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u16_z))) +svuint16_t svmul_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s8_z))) +svint8_t svmul_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s32_z))) +svint32_t svmul_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s64_z))) +svint64_t svmul_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s16_z))) +svint16_t svmul_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f64_m))) +svfloat64_t svmul_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f32_m))) +svfloat32_t svmul_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f16_m))) +svfloat16_t svmul_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f64_x))) +svfloat64_t svmul_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f32_x))) +svfloat32_t svmul_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f16_x))) +svfloat16_t svmul_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f64_z))) +svfloat64_t svmul_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f32_z))) +svfloat32_t svmul_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f16_z))) +svfloat16_t svmul_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u8_m))) +svuint8_t svmul_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u32_m))) +svuint32_t svmul_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u64_m))) +svuint64_t svmul_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u16_m))) +svuint16_t svmul_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s8_m))) +svint8_t svmul_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s32_m))) +svint32_t svmul_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s64_m))) +svint64_t svmul_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s16_m))) +svint16_t svmul_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u8_x))) +svuint8_t svmul_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u32_x))) +svuint32_t svmul_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u64_x))) +svuint64_t svmul_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u16_x))) +svuint16_t svmul_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s8_x))) +svint8_t svmul_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s32_x))) +svint32_t svmul_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s64_x))) +svint64_t svmul_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s16_x))) +svint16_t svmul_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u8_z))) +svuint8_t svmul_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u32_z))) +svuint32_t svmul_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u64_z))) +svuint64_t svmul_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u16_z))) +svuint16_t svmul_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s8_z))) +svint8_t svmul_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s32_z))) +svint32_t svmul_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s64_z))) +svint64_t svmul_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s16_z))) +svint16_t svmul_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_f64))) +svfloat64_t svmul_lane_f64(svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_f32))) +svfloat32_t svmul_lane_f32(svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_f16))) +svfloat16_t svmul_lane_f16(svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s8_m))) +svint8_t svmulh_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s32_m))) +svint32_t svmulh_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s64_m))) +svint64_t svmulh_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s16_m))) +svint16_t svmulh_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s8_x))) +svint8_t svmulh_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s32_x))) +svint32_t svmulh_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s64_x))) +svint64_t svmulh_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s16_x))) +svint16_t svmulh_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s8_z))) +svint8_t svmulh_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s32_z))) +svint32_t svmulh_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s64_z))) +svint64_t svmulh_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s16_z))) +svint16_t svmulh_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u8_m))) +svuint8_t svmulh_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u32_m))) +svuint32_t svmulh_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u64_m))) +svuint64_t svmulh_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u16_m))) +svuint16_t svmulh_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u8_x))) +svuint8_t svmulh_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u32_x))) +svuint32_t svmulh_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u64_x))) +svuint64_t svmulh_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u16_x))) +svuint16_t svmulh_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u8_z))) +svuint8_t svmulh_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u32_z))) +svuint32_t svmulh_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u64_z))) +svuint64_t svmulh_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u16_z))) +svuint16_t svmulh_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s8_m))) +svint8_t svmulh_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s32_m))) +svint32_t svmulh_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s64_m))) +svint64_t svmulh_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s16_m))) +svint16_t svmulh_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s8_x))) +svint8_t svmulh_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s32_x))) +svint32_t svmulh_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s64_x))) +svint64_t svmulh_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s16_x))) +svint16_t svmulh_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s8_z))) +svint8_t svmulh_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s32_z))) +svint32_t svmulh_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s64_z))) +svint64_t svmulh_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s16_z))) +svint16_t svmulh_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u8_m))) +svuint8_t svmulh_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u32_m))) +svuint32_t svmulh_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u64_m))) +svuint64_t svmulh_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u16_m))) +svuint16_t svmulh_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u8_x))) +svuint8_t svmulh_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u32_x))) +svuint32_t svmulh_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u64_x))) +svuint64_t svmulh_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u16_x))) +svuint16_t svmulh_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u8_z))) +svuint8_t svmulh_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u32_z))) +svuint32_t svmulh_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u64_z))) +svuint64_t svmulh_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u16_z))) +svuint16_t svmulh_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f64_m))) +svfloat64_t svmulx_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f32_m))) +svfloat32_t svmulx_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f16_m))) +svfloat16_t svmulx_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f64_x))) +svfloat64_t svmulx_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f32_x))) +svfloat32_t svmulx_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f16_x))) +svfloat16_t svmulx_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f64_z))) +svfloat64_t svmulx_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f32_z))) +svfloat32_t svmulx_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f16_z))) +svfloat16_t svmulx_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f64_m))) +svfloat64_t svmulx_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f32_m))) +svfloat32_t svmulx_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f16_m))) +svfloat16_t svmulx_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f64_x))) +svfloat64_t svmulx_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f32_x))) +svfloat32_t svmulx_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f16_x))) +svfloat16_t svmulx_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f64_z))) +svfloat64_t svmulx_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f32_z))) +svfloat32_t svmulx_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f16_z))) +svfloat16_t svmulx_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnand_b_z))) +svbool_t svnand_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f64_m))) +svfloat64_t svneg_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f32_m))) +svfloat32_t svneg_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f16_m))) +svfloat16_t svneg_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f64_x))) +svfloat64_t svneg_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f32_x))) +svfloat32_t svneg_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f16_x))) +svfloat16_t svneg_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f64_z))) +svfloat64_t svneg_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f32_z))) +svfloat32_t svneg_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f16_z))) +svfloat16_t svneg_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s8_m))) +svint8_t svneg_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s32_m))) +svint32_t svneg_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s64_m))) +svint64_t svneg_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s16_m))) +svint16_t svneg_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s8_x))) +svint8_t svneg_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s32_x))) +svint32_t svneg_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s64_x))) +svint64_t svneg_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s16_x))) +svint16_t svneg_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s8_z))) +svint8_t svneg_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s32_z))) +svint32_t svneg_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s64_z))) +svint64_t svneg_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s16_z))) +svint16_t svneg_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f64_m))) +svfloat64_t svnmad_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f32_m))) +svfloat32_t svnmad_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f16_m))) +svfloat16_t svnmad_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f64_x))) +svfloat64_t svnmad_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f32_x))) +svfloat32_t svnmad_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f16_x))) +svfloat16_t svnmad_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f64_z))) +svfloat64_t svnmad_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f32_z))) +svfloat32_t svnmad_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f16_z))) +svfloat16_t svnmad_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f64_m))) +svfloat64_t svnmad_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f32_m))) +svfloat32_t svnmad_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f16_m))) +svfloat16_t svnmad_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f64_x))) +svfloat64_t svnmad_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f32_x))) +svfloat32_t svnmad_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f16_x))) +svfloat16_t svnmad_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f64_z))) +svfloat64_t svnmad_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f32_z))) +svfloat32_t svnmad_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f16_z))) +svfloat16_t svnmad_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f64_m))) +svfloat64_t svnmla_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f32_m))) +svfloat32_t svnmla_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f16_m))) +svfloat16_t svnmla_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f64_x))) +svfloat64_t svnmla_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f32_x))) +svfloat32_t svnmla_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f16_x))) +svfloat16_t svnmla_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f64_z))) +svfloat64_t svnmla_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f32_z))) +svfloat32_t svnmla_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f16_z))) +svfloat16_t svnmla_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f64_m))) +svfloat64_t svnmla_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f32_m))) +svfloat32_t svnmla_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f16_m))) +svfloat16_t svnmla_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f64_x))) +svfloat64_t svnmla_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f32_x))) +svfloat32_t svnmla_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f16_x))) +svfloat16_t svnmla_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f64_z))) +svfloat64_t svnmla_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f32_z))) +svfloat32_t svnmla_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f16_z))) +svfloat16_t svnmla_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f64_m))) +svfloat64_t svnmls_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f32_m))) +svfloat32_t svnmls_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f16_m))) +svfloat16_t svnmls_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f64_x))) +svfloat64_t svnmls_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f32_x))) +svfloat32_t svnmls_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f16_x))) +svfloat16_t svnmls_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f64_z))) +svfloat64_t svnmls_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f32_z))) +svfloat32_t svnmls_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f16_z))) +svfloat16_t svnmls_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f64_m))) +svfloat64_t svnmls_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f32_m))) +svfloat32_t svnmls_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f16_m))) +svfloat16_t svnmls_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f64_x))) +svfloat64_t svnmls_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f32_x))) +svfloat32_t svnmls_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f16_x))) +svfloat16_t svnmls_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f64_z))) +svfloat64_t svnmls_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f32_z))) +svfloat32_t svnmls_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f16_z))) +svfloat16_t svnmls_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f64_m))) +svfloat64_t svnmsb_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f32_m))) +svfloat32_t svnmsb_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f16_m))) +svfloat16_t svnmsb_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f64_x))) +svfloat64_t svnmsb_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f32_x))) +svfloat32_t svnmsb_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f16_x))) +svfloat16_t svnmsb_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f64_z))) +svfloat64_t svnmsb_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f32_z))) +svfloat32_t svnmsb_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f16_z))) +svfloat16_t svnmsb_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f64_m))) +svfloat64_t svnmsb_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f32_m))) +svfloat32_t svnmsb_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f16_m))) +svfloat16_t svnmsb_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f64_x))) +svfloat64_t svnmsb_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f32_x))) +svfloat32_t svnmsb_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f16_x))) +svfloat16_t svnmsb_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f64_z))) +svfloat64_t svnmsb_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f32_z))) +svfloat32_t svnmsb_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f16_z))) +svfloat16_t svnmsb_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnor_b_z))) +svbool_t svnor_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_b_z))) +svbool_t svnot_b_z(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u8_m))) +svuint8_t svnot_u8_m(svuint8_t, svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u32_m))) +svuint32_t svnot_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u64_m))) +svuint64_t svnot_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u16_m))) +svuint16_t svnot_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s8_m))) +svint8_t svnot_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s32_m))) +svint32_t svnot_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s64_m))) +svint64_t svnot_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s16_m))) +svint16_t svnot_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u8_x))) +svuint8_t svnot_u8_x(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u32_x))) +svuint32_t svnot_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u64_x))) +svuint64_t svnot_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u16_x))) +svuint16_t svnot_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s8_x))) +svint8_t svnot_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s32_x))) +svint32_t svnot_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s64_x))) +svint64_t svnot_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s16_x))) +svint16_t svnot_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u8_z))) +svuint8_t svnot_u8_z(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u32_z))) +svuint32_t svnot_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u64_z))) +svuint64_t svnot_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u16_z))) +svuint16_t svnot_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s8_z))) +svint8_t svnot_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s32_z))) +svint32_t svnot_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s64_z))) +svint64_t svnot_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s16_z))) +svint16_t svnot_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorn_b_z))) +svbool_t svorn_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_b_z))) +svbool_t svorr_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u8_m))) +svuint8_t svorr_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u32_m))) +svuint32_t svorr_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u64_m))) +svuint64_t svorr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u16_m))) +svuint16_t svorr_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s8_m))) +svint8_t svorr_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s32_m))) +svint32_t svorr_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s64_m))) +svint64_t svorr_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s16_m))) +svint16_t svorr_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u8_x))) +svuint8_t svorr_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u32_x))) +svuint32_t svorr_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u64_x))) +svuint64_t svorr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u16_x))) +svuint16_t svorr_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s8_x))) +svint8_t svorr_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s32_x))) +svint32_t svorr_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s64_x))) +svint64_t svorr_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s16_x))) +svint16_t svorr_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u8_z))) +svuint8_t svorr_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u32_z))) +svuint32_t svorr_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u64_z))) +svuint64_t svorr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u16_z))) +svuint16_t svorr_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s8_z))) +svint8_t svorr_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s32_z))) +svint32_t svorr_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s64_z))) +svint64_t svorr_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s16_z))) +svint16_t svorr_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u8_m))) +svuint8_t svorr_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u32_m))) +svuint32_t svorr_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u64_m))) +svuint64_t svorr_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u16_m))) +svuint16_t svorr_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s8_m))) +svint8_t svorr_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s32_m))) +svint32_t svorr_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s64_m))) +svint64_t svorr_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s16_m))) +svint16_t svorr_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u8_x))) +svuint8_t svorr_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u32_x))) +svuint32_t svorr_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u64_x))) +svuint64_t svorr_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u16_x))) +svuint16_t svorr_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s8_x))) +svint8_t svorr_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s32_x))) +svint32_t svorr_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s64_x))) +svint64_t svorr_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s16_x))) +svint16_t svorr_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u8_z))) +svuint8_t svorr_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u32_z))) +svuint32_t svorr_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u64_z))) +svuint64_t svorr_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u16_z))) +svuint16_t svorr_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s8_z))) +svint8_t svorr_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s32_z))) +svint32_t svorr_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s64_z))) +svint64_t svorr_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s16_z))) +svint16_t svorr_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_u8))) +uint8_t svorv_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_u32))) +uint32_t svorv_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_u64))) +uint64_t svorv_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_u16))) +uint16_t svorv_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_s8))) +int8_t svorv_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_s32))) +int32_t svorv_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_s64))) +int64_t svorv_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_s16))) +int16_t svorv_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpfalse_b))) +svbool_t svpfalse_b(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpfirst_b))) +svbool_t svpfirst_b(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpnext_b8))) +svbool_t svpnext_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpnext_b32))) +svbool_t svpnext_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpnext_b64))) +svbool_t svpnext_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpnext_b16))) +svbool_t svpnext_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb))) +void svprfb(svbool_t, void const *, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u32base))) +void svprfb_gather_u32base(svbool_t, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u64base))) +void svprfb_gather_u64base(svbool_t, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u32base_offset))) +void svprfb_gather_u32base_offset(svbool_t, svuint32_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u64base_offset))) +void svprfb_gather_u64base_offset(svbool_t, svuint64_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_s32offset))) +void svprfb_gather_s32offset(svbool_t, void const *, svint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u32offset))) +void svprfb_gather_u32offset(svbool_t, void const *, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_s64offset))) +void svprfb_gather_s64offset(svbool_t, void const *, svint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u64offset))) +void svprfb_gather_u64offset(svbool_t, void const *, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_vnum))) +void svprfb_vnum(svbool_t, void const *, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd))) +void svprfd(svbool_t, void const *, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_u32base))) +void svprfd_gather_u32base(svbool_t, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_u64base))) +void svprfd_gather_u64base(svbool_t, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_u32base_index))) +void svprfd_gather_u32base_index(svbool_t, svuint32_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_u64base_index))) +void svprfd_gather_u64base_index(svbool_t, svuint64_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_s32index))) +void svprfd_gather_s32index(svbool_t, void const *, svint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_u32index))) +void svprfd_gather_u32index(svbool_t, void const *, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_s64index))) +void svprfd_gather_s64index(svbool_t, void const *, svint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_u64index))) +void svprfd_gather_u64index(svbool_t, void const *, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_vnum))) +void svprfd_vnum(svbool_t, void const *, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh))) +void svprfh(svbool_t, void const *, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_u32base))) +void svprfh_gather_u32base(svbool_t, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_u64base))) +void svprfh_gather_u64base(svbool_t, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_u32base_index))) +void svprfh_gather_u32base_index(svbool_t, svuint32_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_u64base_index))) +void svprfh_gather_u64base_index(svbool_t, svuint64_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_s32index))) +void svprfh_gather_s32index(svbool_t, void const *, svint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_u32index))) +void svprfh_gather_u32index(svbool_t, void const *, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_s64index))) +void svprfh_gather_s64index(svbool_t, void const *, svint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_u64index))) +void svprfh_gather_u64index(svbool_t, void const *, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_vnum))) +void svprfh_vnum(svbool_t, void const *, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw))) +void svprfw(svbool_t, void const *, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_u32base))) +void svprfw_gather_u32base(svbool_t, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_u64base))) +void svprfw_gather_u64base(svbool_t, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_u32base_index))) +void svprfw_gather_u32base_index(svbool_t, svuint32_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_u64base_index))) +void svprfw_gather_u64base_index(svbool_t, svuint64_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_s32index))) +void svprfw_gather_s32index(svbool_t, void const *, svint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_u32index))) +void svprfw_gather_u32index(svbool_t, void const *, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_s64index))) +void svprfw_gather_s64index(svbool_t, void const *, svint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_u64index))) +void svprfw_gather_u64index(svbool_t, void const *, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_vnum))) +void svprfw_vnum(svbool_t, void const *, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptest_any))) +bool svptest_any(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptest_first))) +bool svptest_first(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptest_last))) +bool svptest_last(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_pat_b8))) +svbool_t svptrue_pat_b8(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_pat_b32))) +svbool_t svptrue_pat_b32(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_pat_b64))) +svbool_t svptrue_pat_b64(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_pat_b16))) +svbool_t svptrue_pat_b16(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_b8))) +svbool_t svptrue_b8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_b32))) +svbool_t svptrue_b32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_b64))) +svbool_t svptrue_b64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_b16))) +svbool_t svptrue_b16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s8))) +svint8_t svqadd_n_s8(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s32))) +svint32_t svqadd_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s64))) +svint64_t svqadd_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s16))) +svint16_t svqadd_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u8))) +svuint8_t svqadd_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u32))) +svuint32_t svqadd_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u64))) +svuint64_t svqadd_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u16))) +svuint16_t svqadd_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s8))) +svint8_t svqadd_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s32))) +svint32_t svqadd_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s64))) +svint64_t svqadd_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s16))) +svint16_t svqadd_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u8))) +svuint8_t svqadd_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u32))) +svuint32_t svqadd_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u64))) +svuint64_t svqadd_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u16))) +svuint16_t svqadd_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_n_s32))) +int32_t svqdecb_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_n_s64))) +int64_t svqdecb_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_n_u32))) +uint32_t svqdecb_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_n_u64))) +uint64_t svqdecb_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_pat_n_s32))) +int32_t svqdecb_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_pat_n_s64))) +int64_t svqdecb_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_pat_n_u32))) +uint32_t svqdecb_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_pat_n_u64))) +uint64_t svqdecb_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_n_s32))) +int32_t svqdecd_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_n_s64))) +int64_t svqdecd_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_n_u32))) +uint32_t svqdecd_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_n_u64))) +uint64_t svqdecd_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_s64))) +svint64_t svqdecd_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_u64))) +svuint64_t svqdecd_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_pat_n_s32))) +int32_t svqdecd_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_pat_n_s64))) +int64_t svqdecd_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_pat_n_u32))) +uint32_t svqdecd_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_pat_n_u64))) +uint64_t svqdecd_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_pat_s64))) +svint64_t svqdecd_pat_s64(svint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_pat_u64))) +svuint64_t svqdecd_pat_u64(svuint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_n_s32))) +int32_t svqdech_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_n_s64))) +int64_t svqdech_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_n_u32))) +uint32_t svqdech_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_n_u64))) +uint64_t svqdech_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_s16))) +svint16_t svqdech_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_u16))) +svuint16_t svqdech_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_pat_n_s32))) +int32_t svqdech_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_pat_n_s64))) +int64_t svqdech_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_pat_n_u32))) +uint32_t svqdech_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_pat_n_u64))) +uint64_t svqdech_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_pat_s16))) +svint16_t svqdech_pat_s16(svint16_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_pat_u16))) +svuint16_t svqdech_pat_u16(svuint16_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s32_b8))) +int32_t svqdecp_n_s32_b8(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s32_b32))) +int32_t svqdecp_n_s32_b32(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s32_b64))) +int32_t svqdecp_n_s32_b64(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s32_b16))) +int32_t svqdecp_n_s32_b16(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s64_b8))) +int64_t svqdecp_n_s64_b8(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s64_b32))) +int64_t svqdecp_n_s64_b32(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s64_b64))) +int64_t svqdecp_n_s64_b64(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s64_b16))) +int64_t svqdecp_n_s64_b16(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u32_b8))) +uint32_t svqdecp_n_u32_b8(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u32_b32))) +uint32_t svqdecp_n_u32_b32(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u32_b64))) +uint32_t svqdecp_n_u32_b64(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u32_b16))) +uint32_t svqdecp_n_u32_b16(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u64_b8))) +uint64_t svqdecp_n_u64_b8(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u64_b32))) +uint64_t svqdecp_n_u64_b32(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u64_b64))) +uint64_t svqdecp_n_u64_b64(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u64_b16))) +uint64_t svqdecp_n_u64_b16(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_s32))) +svint32_t svqdecp_s32(svint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_s64))) +svint64_t svqdecp_s64(svint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_s16))) +svint16_t svqdecp_s16(svint16_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_u32))) +svuint32_t svqdecp_u32(svuint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_u64))) +svuint64_t svqdecp_u64(svuint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_u16))) +svuint16_t svqdecp_u16(svuint16_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_n_s32))) +int32_t svqdecw_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_n_s64))) +int64_t svqdecw_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_n_u32))) +uint32_t svqdecw_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_n_u64))) +uint64_t svqdecw_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_s32))) +svint32_t svqdecw_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_u32))) +svuint32_t svqdecw_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_pat_n_s32))) +int32_t svqdecw_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_pat_n_s64))) +int64_t svqdecw_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_pat_n_u32))) +uint32_t svqdecw_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_pat_n_u64))) +uint64_t svqdecw_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_pat_s32))) +svint32_t svqdecw_pat_s32(svint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_pat_u32))) +svuint32_t svqdecw_pat_u32(svuint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_n_s32))) +int32_t svqincb_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_n_s64))) +int64_t svqincb_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_n_u32))) +uint32_t svqincb_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_n_u64))) +uint64_t svqincb_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_pat_n_s32))) +int32_t svqincb_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_pat_n_s64))) +int64_t svqincb_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_pat_n_u32))) +uint32_t svqincb_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_pat_n_u64))) +uint64_t svqincb_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_n_s32))) +int32_t svqincd_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_n_s64))) +int64_t svqincd_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_n_u32))) +uint32_t svqincd_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_n_u64))) +uint64_t svqincd_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_s64))) +svint64_t svqincd_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_u64))) +svuint64_t svqincd_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_pat_n_s32))) +int32_t svqincd_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_pat_n_s64))) +int64_t svqincd_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_pat_n_u32))) +uint32_t svqincd_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_pat_n_u64))) +uint64_t svqincd_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_pat_s64))) +svint64_t svqincd_pat_s64(svint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_pat_u64))) +svuint64_t svqincd_pat_u64(svuint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_n_s32))) +int32_t svqinch_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_n_s64))) +int64_t svqinch_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_n_u32))) +uint32_t svqinch_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_n_u64))) +uint64_t svqinch_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_s16))) +svint16_t svqinch_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_u16))) +svuint16_t svqinch_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_pat_n_s32))) +int32_t svqinch_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_pat_n_s64))) +int64_t svqinch_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_pat_n_u32))) +uint32_t svqinch_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_pat_n_u64))) +uint64_t svqinch_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_pat_s16))) +svint16_t svqinch_pat_s16(svint16_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_pat_u16))) +svuint16_t svqinch_pat_u16(svuint16_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s32_b8))) +int32_t svqincp_n_s32_b8(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s32_b32))) +int32_t svqincp_n_s32_b32(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s32_b64))) +int32_t svqincp_n_s32_b64(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s32_b16))) +int32_t svqincp_n_s32_b16(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s64_b8))) +int64_t svqincp_n_s64_b8(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s64_b32))) +int64_t svqincp_n_s64_b32(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s64_b64))) +int64_t svqincp_n_s64_b64(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s64_b16))) +int64_t svqincp_n_s64_b16(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u32_b8))) +uint32_t svqincp_n_u32_b8(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u32_b32))) +uint32_t svqincp_n_u32_b32(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u32_b64))) +uint32_t svqincp_n_u32_b64(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u32_b16))) +uint32_t svqincp_n_u32_b16(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u64_b8))) +uint64_t svqincp_n_u64_b8(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u64_b32))) +uint64_t svqincp_n_u64_b32(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u64_b64))) +uint64_t svqincp_n_u64_b64(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u64_b16))) +uint64_t svqincp_n_u64_b16(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_s32))) +svint32_t svqincp_s32(svint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_s64))) +svint64_t svqincp_s64(svint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_s16))) +svint16_t svqincp_s16(svint16_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_u32))) +svuint32_t svqincp_u32(svuint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_u64))) +svuint64_t svqincp_u64(svuint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_u16))) +svuint16_t svqincp_u16(svuint16_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_n_s32))) +int32_t svqincw_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_n_s64))) +int64_t svqincw_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_n_u32))) +uint32_t svqincw_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_n_u64))) +uint64_t svqincw_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_s32))) +svint32_t svqincw_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_u32))) +svuint32_t svqincw_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_pat_n_s32))) +int32_t svqincw_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_pat_n_s64))) +int64_t svqincw_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_pat_n_u32))) +uint32_t svqincw_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_pat_n_u64))) +uint64_t svqincw_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_pat_s32))) +svint32_t svqincw_pat_s32(svint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_pat_u32))) +svuint32_t svqincw_pat_u32(svuint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s8))) +svint8_t svqsub_n_s8(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s32))) +svint32_t svqsub_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s64))) +svint64_t svqsub_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s16))) +svint16_t svqsub_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u8))) +svuint8_t svqsub_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u32))) +svuint32_t svqsub_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u64))) +svuint64_t svqsub_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u16))) +svuint16_t svqsub_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s8))) +svint8_t svqsub_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s32))) +svint32_t svqsub_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s64))) +svint64_t svqsub_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s16))) +svint16_t svqsub_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u8))) +svuint8_t svqsub_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u32))) +svuint32_t svqsub_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u64))) +svuint64_t svqsub_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u16))) +svuint16_t svqsub_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u8_m))) +svuint8_t svrbit_u8_m(svuint8_t, svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u32_m))) +svuint32_t svrbit_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u64_m))) +svuint64_t svrbit_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u16_m))) +svuint16_t svrbit_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s8_m))) +svint8_t svrbit_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s32_m))) +svint32_t svrbit_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s64_m))) +svint64_t svrbit_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s16_m))) +svint16_t svrbit_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u8_x))) +svuint8_t svrbit_u8_x(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u32_x))) +svuint32_t svrbit_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u64_x))) +svuint64_t svrbit_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u16_x))) +svuint16_t svrbit_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s8_x))) +svint8_t svrbit_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s32_x))) +svint32_t svrbit_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s64_x))) +svint64_t svrbit_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s16_x))) +svint16_t svrbit_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u8_z))) +svuint8_t svrbit_u8_z(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u32_z))) +svuint32_t svrbit_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u64_z))) +svuint64_t svrbit_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u16_z))) +svuint16_t svrbit_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s8_z))) +svint8_t svrbit_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s32_z))) +svint32_t svrbit_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s64_z))) +svint64_t svrbit_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s16_z))) +svint16_t svrbit_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrdffr))) +svbool_t svrdffr(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrdffr_z))) +svbool_t svrdffr_z(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpe_f64))) +svfloat64_t svrecpe_f64(svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpe_f32))) +svfloat32_t svrecpe_f32(svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpe_f16))) +svfloat16_t svrecpe_f16(svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecps_f64))) +svfloat64_t svrecps_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecps_f32))) +svfloat32_t svrecps_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecps_f16))) +svfloat16_t svrecps_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f64_m))) +svfloat64_t svrecpx_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f32_m))) +svfloat32_t svrecpx_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f16_m))) +svfloat16_t svrecpx_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f64_x))) +svfloat64_t svrecpx_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f32_x))) +svfloat32_t svrecpx_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f16_x))) +svfloat16_t svrecpx_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f64_z))) +svfloat64_t svrecpx_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f32_z))) +svfloat32_t svrecpx_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f16_z))) +svfloat16_t svrecpx_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_u8))) +svuint8_t svrev_u8(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_u32))) +svuint32_t svrev_u32(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_u64))) +svuint64_t svrev_u64(svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_u16))) +svuint16_t svrev_u16(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_s8))) +svint8_t svrev_s8(svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_f64))) +svfloat64_t svrev_f64(svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_f32))) +svfloat32_t svrev_f32(svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_f16))) +svfloat16_t svrev_f16(svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_s32))) +svint32_t svrev_s32(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_s64))) +svint64_t svrev_s64(svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_s16))) +svint16_t svrev_s16(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_b8))) +svbool_t svrev_b8(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_b32))) +svbool_t svrev_b32(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_b64))) +svbool_t svrev_b64(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_b16))) +svbool_t svrev_b16(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u32_m))) +svuint32_t svrevb_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u64_m))) +svuint64_t svrevb_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u16_m))) +svuint16_t svrevb_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s32_m))) +svint32_t svrevb_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s64_m))) +svint64_t svrevb_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s16_m))) +svint16_t svrevb_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u32_x))) +svuint32_t svrevb_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u64_x))) +svuint64_t svrevb_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u16_x))) +svuint16_t svrevb_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s32_x))) +svint32_t svrevb_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s64_x))) +svint64_t svrevb_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s16_x))) +svint16_t svrevb_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u32_z))) +svuint32_t svrevb_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u64_z))) +svuint64_t svrevb_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u16_z))) +svuint16_t svrevb_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s32_z))) +svint32_t svrevb_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s64_z))) +svint64_t svrevb_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s16_z))) +svint16_t svrevb_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_u32_m))) +svuint32_t svrevh_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_u64_m))) +svuint64_t svrevh_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_s32_m))) +svint32_t svrevh_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_s64_m))) +svint64_t svrevh_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_u32_x))) +svuint32_t svrevh_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_u64_x))) +svuint64_t svrevh_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_s32_x))) +svint32_t svrevh_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_s64_x))) +svint64_t svrevh_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_u32_z))) +svuint32_t svrevh_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_u64_z))) +svuint64_t svrevh_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_s32_z))) +svint32_t svrevh_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_s64_z))) +svint64_t svrevh_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevw_u64_m))) +svuint64_t svrevw_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevw_s64_m))) +svint64_t svrevw_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevw_u64_x))) +svuint64_t svrevw_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevw_s64_x))) +svint64_t svrevw_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevw_u64_z))) +svuint64_t svrevw_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevw_s64_z))) +svint64_t svrevw_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f64_m))) +svfloat64_t svrinta_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f32_m))) +svfloat32_t svrinta_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f16_m))) +svfloat16_t svrinta_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f64_x))) +svfloat64_t svrinta_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f32_x))) +svfloat32_t svrinta_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f16_x))) +svfloat16_t svrinta_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f64_z))) +svfloat64_t svrinta_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f32_z))) +svfloat32_t svrinta_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f16_z))) +svfloat16_t svrinta_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f64_m))) +svfloat64_t svrinti_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f32_m))) +svfloat32_t svrinti_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f16_m))) +svfloat16_t svrinti_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f64_x))) +svfloat64_t svrinti_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f32_x))) +svfloat32_t svrinti_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f16_x))) +svfloat16_t svrinti_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f64_z))) +svfloat64_t svrinti_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f32_z))) +svfloat32_t svrinti_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f16_z))) +svfloat16_t svrinti_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f64_m))) +svfloat64_t svrintm_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f32_m))) +svfloat32_t svrintm_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f16_m))) +svfloat16_t svrintm_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f64_x))) +svfloat64_t svrintm_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f32_x))) +svfloat32_t svrintm_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f16_x))) +svfloat16_t svrintm_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f64_z))) +svfloat64_t svrintm_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f32_z))) +svfloat32_t svrintm_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f16_z))) +svfloat16_t svrintm_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f64_m))) +svfloat64_t svrintn_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f32_m))) +svfloat32_t svrintn_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f16_m))) +svfloat16_t svrintn_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f64_x))) +svfloat64_t svrintn_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f32_x))) +svfloat32_t svrintn_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f16_x))) +svfloat16_t svrintn_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f64_z))) +svfloat64_t svrintn_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f32_z))) +svfloat32_t svrintn_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f16_z))) +svfloat16_t svrintn_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f64_m))) +svfloat64_t svrintp_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f32_m))) +svfloat32_t svrintp_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f16_m))) +svfloat16_t svrintp_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f64_x))) +svfloat64_t svrintp_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f32_x))) +svfloat32_t svrintp_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f16_x))) +svfloat16_t svrintp_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f64_z))) +svfloat64_t svrintp_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f32_z))) +svfloat32_t svrintp_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f16_z))) +svfloat16_t svrintp_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f64_m))) +svfloat64_t svrintx_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f32_m))) +svfloat32_t svrintx_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f16_m))) +svfloat16_t svrintx_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f64_x))) +svfloat64_t svrintx_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f32_x))) +svfloat32_t svrintx_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f16_x))) +svfloat16_t svrintx_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f64_z))) +svfloat64_t svrintx_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f32_z))) +svfloat32_t svrintx_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f16_z))) +svfloat16_t svrintx_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f64_m))) +svfloat64_t svrintz_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f32_m))) +svfloat32_t svrintz_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f16_m))) +svfloat16_t svrintz_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f64_x))) +svfloat64_t svrintz_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f32_x))) +svfloat32_t svrintz_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f16_x))) +svfloat16_t svrintz_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f64_z))) +svfloat64_t svrintz_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f32_z))) +svfloat32_t svrintz_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f16_z))) +svfloat16_t svrintz_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrte_f64))) +svfloat64_t svrsqrte_f64(svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrte_f32))) +svfloat32_t svrsqrte_f32(svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrte_f16))) +svfloat16_t svrsqrte_f16(svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrts_f64))) +svfloat64_t svrsqrts_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrts_f32))) +svfloat32_t svrsqrts_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrts_f16))) +svfloat16_t svrsqrts_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f64_m))) +svfloat64_t svscale_n_f64_m(svbool_t, svfloat64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f32_m))) +svfloat32_t svscale_n_f32_m(svbool_t, svfloat32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f16_m))) +svfloat16_t svscale_n_f16_m(svbool_t, svfloat16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f64_x))) +svfloat64_t svscale_n_f64_x(svbool_t, svfloat64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f32_x))) +svfloat32_t svscale_n_f32_x(svbool_t, svfloat32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f16_x))) +svfloat16_t svscale_n_f16_x(svbool_t, svfloat16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f64_z))) +svfloat64_t svscale_n_f64_z(svbool_t, svfloat64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f32_z))) +svfloat32_t svscale_n_f32_z(svbool_t, svfloat32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f16_z))) +svfloat16_t svscale_n_f16_z(svbool_t, svfloat16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f64_m))) +svfloat64_t svscale_f64_m(svbool_t, svfloat64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f32_m))) +svfloat32_t svscale_f32_m(svbool_t, svfloat32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f16_m))) +svfloat16_t svscale_f16_m(svbool_t, svfloat16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f64_x))) +svfloat64_t svscale_f64_x(svbool_t, svfloat64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f32_x))) +svfloat32_t svscale_f32_x(svbool_t, svfloat32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f16_x))) +svfloat16_t svscale_f16_x(svbool_t, svfloat16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f64_z))) +svfloat64_t svscale_f64_z(svbool_t, svfloat64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f32_z))) +svfloat32_t svscale_f32_z(svbool_t, svfloat32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f16_z))) +svfloat16_t svscale_f16_z(svbool_t, svfloat16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_b))) +svbool_t svsel_b(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_u8))) +svuint8_t svsel_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_u32))) +svuint32_t svsel_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_u64))) +svuint64_t svsel_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_u16))) +svuint16_t svsel_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_s8))) +svint8_t svsel_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_f64))) +svfloat64_t svsel_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_f32))) +svfloat32_t svsel_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_f16))) +svfloat16_t svsel_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_s32))) +svint32_t svsel_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_s64))) +svint64_t svsel_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_s16))) +svint16_t svsel_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_u8))) +svuint8x2_t svset2_u8(svuint8x2_t, uint64_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_u32))) +svuint32x2_t svset2_u32(svuint32x2_t, uint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_u64))) +svuint64x2_t svset2_u64(svuint64x2_t, uint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_u16))) +svuint16x2_t svset2_u16(svuint16x2_t, uint64_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_s8))) +svint8x2_t svset2_s8(svint8x2_t, uint64_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_f64))) +svfloat64x2_t svset2_f64(svfloat64x2_t, uint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_f32))) +svfloat32x2_t svset2_f32(svfloat32x2_t, uint64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_f16))) +svfloat16x2_t svset2_f16(svfloat16x2_t, uint64_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_s32))) +svint32x2_t svset2_s32(svint32x2_t, uint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_s64))) +svint64x2_t svset2_s64(svint64x2_t, uint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_s16))) +svint16x2_t svset2_s16(svint16x2_t, uint64_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_u8))) +svuint8x3_t svset3_u8(svuint8x3_t, uint64_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_u32))) +svuint32x3_t svset3_u32(svuint32x3_t, uint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_u64))) +svuint64x3_t svset3_u64(svuint64x3_t, uint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_u16))) +svuint16x3_t svset3_u16(svuint16x3_t, uint64_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_s8))) +svint8x3_t svset3_s8(svint8x3_t, uint64_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_f64))) +svfloat64x3_t svset3_f64(svfloat64x3_t, uint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_f32))) +svfloat32x3_t svset3_f32(svfloat32x3_t, uint64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_f16))) +svfloat16x3_t svset3_f16(svfloat16x3_t, uint64_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_s32))) +svint32x3_t svset3_s32(svint32x3_t, uint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_s64))) +svint64x3_t svset3_s64(svint64x3_t, uint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_s16))) +svint16x3_t svset3_s16(svint16x3_t, uint64_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_u8))) +svuint8x4_t svset4_u8(svuint8x4_t, uint64_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_u32))) +svuint32x4_t svset4_u32(svuint32x4_t, uint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_u64))) +svuint64x4_t svset4_u64(svuint64x4_t, uint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_u16))) +svuint16x4_t svset4_u16(svuint16x4_t, uint64_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_s8))) +svint8x4_t svset4_s8(svint8x4_t, uint64_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_f64))) +svfloat64x4_t svset4_f64(svfloat64x4_t, uint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_f32))) +svfloat32x4_t svset4_f32(svfloat32x4_t, uint64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_f16))) +svfloat16x4_t svset4_f16(svfloat16x4_t, uint64_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_s32))) +svint32x4_t svset4_s32(svint32x4_t, uint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_s64))) +svint64x4_t svset4_s64(svint64x4_t, uint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_s16))) +svint16x4_t svset4_s16(svint16x4_t, uint64_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsetffr))) +void svsetffr(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_u8))) +svuint8_t svsplice_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_u32))) +svuint32_t svsplice_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_u64))) +svuint64_t svsplice_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_u16))) +svuint16_t svsplice_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_s8))) +svint8_t svsplice_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_f64))) +svfloat64_t svsplice_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_f32))) +svfloat32_t svsplice_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_f16))) +svfloat16_t svsplice_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_s32))) +svint32_t svsplice_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_s64))) +svint64_t svsplice_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_s16))) +svint16_t svsplice_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f64_m))) +svfloat64_t svsqrt_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f32_m))) +svfloat32_t svsqrt_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f16_m))) +svfloat16_t svsqrt_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f64_x))) +svfloat64_t svsqrt_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f32_x))) +svfloat32_t svsqrt_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f16_x))) +svfloat16_t svsqrt_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f64_z))) +svfloat64_t svsqrt_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f32_z))) +svfloat32_t svsqrt_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f16_z))) +svfloat16_t svsqrt_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_u8))) +void svst1_u8(svbool_t, uint8_t *, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_u32))) +void svst1_u32(svbool_t, uint32_t *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_u64))) +void svst1_u64(svbool_t, uint64_t *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_u16))) +void svst1_u16(svbool_t, uint16_t *, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_s8))) +void svst1_s8(svbool_t, int8_t *, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_f64))) +void svst1_f64(svbool_t, float64_t *, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_f32))) +void svst1_f32(svbool_t, float32_t *, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_f16))) +void svst1_f16(svbool_t, float16_t *, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_s32))) +void svst1_s32(svbool_t, int32_t *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_s64))) +void svst1_s64(svbool_t, int64_t *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_s16))) +void svst1_s16(svbool_t, int16_t *, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_index_u32))) +void svst1_scatter_u32base_index_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_index_u64))) +void svst1_scatter_u64base_index_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_index_f64))) +void svst1_scatter_u64base_index_f64(svbool_t, svuint64_t, int64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_index_f32))) +void svst1_scatter_u32base_index_f32(svbool_t, svuint32_t, int64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_index_s32))) +void svst1_scatter_u32base_index_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_index_s64))) +void svst1_scatter_u64base_index_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_offset_u32))) +void svst1_scatter_u32base_offset_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_offset_u64))) +void svst1_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_offset_f64))) +void svst1_scatter_u64base_offset_f64(svbool_t, svuint64_t, int64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_offset_f32))) +void svst1_scatter_u32base_offset_f32(svbool_t, svuint32_t, int64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_offset_s32))) +void svst1_scatter_u32base_offset_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_offset_s64))) +void svst1_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_u32))) +void svst1_scatter_u32base_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_u64))) +void svst1_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_f64))) +void svst1_scatter_u64base_f64(svbool_t, svuint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_f32))) +void svst1_scatter_u32base_f32(svbool_t, svuint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_s32))) +void svst1_scatter_u32base_s32(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_s64))) +void svst1_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s32index_u32))) +void svst1_scatter_s32index_u32(svbool_t, uint32_t *, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s32index_f32))) +void svst1_scatter_s32index_f32(svbool_t, float32_t *, svint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s32index_s32))) +void svst1_scatter_s32index_s32(svbool_t, int32_t *, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32index_u32))) +void svst1_scatter_u32index_u32(svbool_t, uint32_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32index_f32))) +void svst1_scatter_u32index_f32(svbool_t, float32_t *, svuint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32index_s32))) +void svst1_scatter_u32index_s32(svbool_t, int32_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s64index_u64))) +void svst1_scatter_s64index_u64(svbool_t, uint64_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s64index_f64))) +void svst1_scatter_s64index_f64(svbool_t, float64_t *, svint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s64index_s64))) +void svst1_scatter_s64index_s64(svbool_t, int64_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64index_u64))) +void svst1_scatter_u64index_u64(svbool_t, uint64_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64index_f64))) +void svst1_scatter_u64index_f64(svbool_t, float64_t *, svuint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64index_s64))) +void svst1_scatter_u64index_s64(svbool_t, int64_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s32offset_u32))) +void svst1_scatter_s32offset_u32(svbool_t, uint32_t *, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s32offset_f32))) +void svst1_scatter_s32offset_f32(svbool_t, float32_t *, svint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s32offset_s32))) +void svst1_scatter_s32offset_s32(svbool_t, int32_t *, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32offset_u32))) +void svst1_scatter_u32offset_u32(svbool_t, uint32_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32offset_f32))) +void svst1_scatter_u32offset_f32(svbool_t, float32_t *, svuint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32offset_s32))) +void svst1_scatter_u32offset_s32(svbool_t, int32_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s64offset_u64))) +void svst1_scatter_s64offset_u64(svbool_t, uint64_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s64offset_f64))) +void svst1_scatter_s64offset_f64(svbool_t, float64_t *, svint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s64offset_s64))) +void svst1_scatter_s64offset_s64(svbool_t, int64_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64offset_u64))) +void svst1_scatter_u64offset_u64(svbool_t, uint64_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64offset_f64))) +void svst1_scatter_u64offset_f64(svbool_t, float64_t *, svuint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64offset_s64))) +void svst1_scatter_u64offset_s64(svbool_t, int64_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_u8))) +void svst1_vnum_u8(svbool_t, uint8_t *, int64_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_u32))) +void svst1_vnum_u32(svbool_t, uint32_t *, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_u64))) +void svst1_vnum_u64(svbool_t, uint64_t *, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_u16))) +void svst1_vnum_u16(svbool_t, uint16_t *, int64_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_s8))) +void svst1_vnum_s8(svbool_t, int8_t *, int64_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_f64))) +void svst1_vnum_f64(svbool_t, float64_t *, int64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_f32))) +void svst1_vnum_f32(svbool_t, float32_t *, int64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_f16))) +void svst1_vnum_f16(svbool_t, float16_t *, int64_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_s32))) +void svst1_vnum_s32(svbool_t, int32_t *, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_s64))) +void svst1_vnum_s64(svbool_t, int64_t *, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_s16))) +void svst1_vnum_s16(svbool_t, int16_t *, int64_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_s32))) +void svst1b_s32(svbool_t, int8_t *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_s64))) +void svst1b_s64(svbool_t, int8_t *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_s16))) +void svst1b_s16(svbool_t, int8_t *, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_u32))) +void svst1b_u32(svbool_t, uint8_t *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_u64))) +void svst1b_u64(svbool_t, uint8_t *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_u16))) +void svst1b_u16(svbool_t, uint8_t *, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u32base_offset_u32))) +void svst1b_scatter_u32base_offset_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u64base_offset_u64))) +void svst1b_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u32base_offset_s32))) +void svst1b_scatter_u32base_offset_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u64base_offset_s64))) +void svst1b_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u32base_u32))) +void svst1b_scatter_u32base_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u64base_u64))) +void svst1b_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u32base_s32))) +void svst1b_scatter_u32base_s32(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u64base_s64))) +void svst1b_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_s32offset_s32))) +void svst1b_scatter_s32offset_s32(svbool_t, int8_t *, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_s32offset_u32))) +void svst1b_scatter_s32offset_u32(svbool_t, uint8_t *, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u32offset_s32))) +void svst1b_scatter_u32offset_s32(svbool_t, int8_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u32offset_u32))) +void svst1b_scatter_u32offset_u32(svbool_t, uint8_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_s64offset_s64))) +void svst1b_scatter_s64offset_s64(svbool_t, int8_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_s64offset_u64))) +void svst1b_scatter_s64offset_u64(svbool_t, uint8_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u64offset_s64))) +void svst1b_scatter_u64offset_s64(svbool_t, int8_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u64offset_u64))) +void svst1b_scatter_u64offset_u64(svbool_t, uint8_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_vnum_s32))) +void svst1b_vnum_s32(svbool_t, int8_t *, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_vnum_s64))) +void svst1b_vnum_s64(svbool_t, int8_t *, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_vnum_s16))) +void svst1b_vnum_s16(svbool_t, int8_t *, int64_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_vnum_u32))) +void svst1b_vnum_u32(svbool_t, uint8_t *, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_vnum_u64))) +void svst1b_vnum_u64(svbool_t, uint8_t *, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_vnum_u16))) +void svst1b_vnum_u16(svbool_t, uint8_t *, int64_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_s32))) +void svst1h_s32(svbool_t, int16_t *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_s64))) +void svst1h_s64(svbool_t, int16_t *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_u32))) +void svst1h_u32(svbool_t, uint16_t *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_u64))) +void svst1h_u64(svbool_t, uint16_t *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32base_index_u32))) +void svst1h_scatter_u32base_index_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64base_index_u64))) +void svst1h_scatter_u64base_index_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32base_index_s32))) +void svst1h_scatter_u32base_index_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64base_index_s64))) +void svst1h_scatter_u64base_index_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32base_offset_u32))) +void svst1h_scatter_u32base_offset_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64base_offset_u64))) +void svst1h_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32base_offset_s32))) +void svst1h_scatter_u32base_offset_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64base_offset_s64))) +void svst1h_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32base_u32))) +void svst1h_scatter_u32base_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64base_u64))) +void svst1h_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32base_s32))) +void svst1h_scatter_u32base_s32(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64base_s64))) +void svst1h_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s32index_s32))) +void svst1h_scatter_s32index_s32(svbool_t, int16_t *, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s32index_u32))) +void svst1h_scatter_s32index_u32(svbool_t, uint16_t *, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32index_s32))) +void svst1h_scatter_u32index_s32(svbool_t, int16_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32index_u32))) +void svst1h_scatter_u32index_u32(svbool_t, uint16_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s64index_s64))) +void svst1h_scatter_s64index_s64(svbool_t, int16_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s64index_u64))) +void svst1h_scatter_s64index_u64(svbool_t, uint16_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64index_s64))) +void svst1h_scatter_u64index_s64(svbool_t, int16_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64index_u64))) +void svst1h_scatter_u64index_u64(svbool_t, uint16_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s32offset_s32))) +void svst1h_scatter_s32offset_s32(svbool_t, int16_t *, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s32offset_u32))) +void svst1h_scatter_s32offset_u32(svbool_t, uint16_t *, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32offset_s32))) +void svst1h_scatter_u32offset_s32(svbool_t, int16_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32offset_u32))) +void svst1h_scatter_u32offset_u32(svbool_t, uint16_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s64offset_s64))) +void svst1h_scatter_s64offset_s64(svbool_t, int16_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s64offset_u64))) +void svst1h_scatter_s64offset_u64(svbool_t, uint16_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64offset_s64))) +void svst1h_scatter_u64offset_s64(svbool_t, int16_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64offset_u64))) +void svst1h_scatter_u64offset_u64(svbool_t, uint16_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_vnum_s32))) +void svst1h_vnum_s32(svbool_t, int16_t *, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_vnum_s64))) +void svst1h_vnum_s64(svbool_t, int16_t *, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_vnum_u32))) +void svst1h_vnum_u32(svbool_t, uint16_t *, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_vnum_u64))) +void svst1h_vnum_u64(svbool_t, uint16_t *, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_s64))) +void svst1w_s64(svbool_t, int32_t *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_u64))) +void svst1w_u64(svbool_t, uint32_t *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64base_index_u64))) +void svst1w_scatter_u64base_index_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64base_index_s64))) +void svst1w_scatter_u64base_index_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64base_offset_u64))) +void svst1w_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64base_offset_s64))) +void svst1w_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64base_u64))) +void svst1w_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64base_s64))) +void svst1w_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_s64index_s64))) +void svst1w_scatter_s64index_s64(svbool_t, int32_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_s64index_u64))) +void svst1w_scatter_s64index_u64(svbool_t, uint32_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64index_s64))) +void svst1w_scatter_u64index_s64(svbool_t, int32_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64index_u64))) +void svst1w_scatter_u64index_u64(svbool_t, uint32_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_s64offset_s64))) +void svst1w_scatter_s64offset_s64(svbool_t, int32_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_s64offset_u64))) +void svst1w_scatter_s64offset_u64(svbool_t, uint32_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64offset_s64))) +void svst1w_scatter_u64offset_s64(svbool_t, int32_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64offset_u64))) +void svst1w_scatter_u64offset_u64(svbool_t, uint32_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_vnum_s64))) +void svst1w_vnum_s64(svbool_t, int32_t *, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_vnum_u64))) +void svst1w_vnum_u64(svbool_t, uint32_t *, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_u8))) +void svst2_u8(svbool_t, uint8_t *, svuint8x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_u32))) +void svst2_u32(svbool_t, uint32_t *, svuint32x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_u64))) +void svst2_u64(svbool_t, uint64_t *, svuint64x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_u16))) +void svst2_u16(svbool_t, uint16_t *, svuint16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_s8))) +void svst2_s8(svbool_t, int8_t *, svint8x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_f64))) +void svst2_f64(svbool_t, float64_t *, svfloat64x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_f32))) +void svst2_f32(svbool_t, float32_t *, svfloat32x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_f16))) +void svst2_f16(svbool_t, float16_t *, svfloat16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_s32))) +void svst2_s32(svbool_t, int32_t *, svint32x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_s64))) +void svst2_s64(svbool_t, int64_t *, svint64x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_s16))) +void svst2_s16(svbool_t, int16_t *, svint16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_u8))) +void svst2_vnum_u8(svbool_t, uint8_t *, int64_t, svuint8x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_u32))) +void svst2_vnum_u32(svbool_t, uint32_t *, int64_t, svuint32x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_u64))) +void svst2_vnum_u64(svbool_t, uint64_t *, int64_t, svuint64x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_u16))) +void svst2_vnum_u16(svbool_t, uint16_t *, int64_t, svuint16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_s8))) +void svst2_vnum_s8(svbool_t, int8_t *, int64_t, svint8x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_f64))) +void svst2_vnum_f64(svbool_t, float64_t *, int64_t, svfloat64x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_f32))) +void svst2_vnum_f32(svbool_t, float32_t *, int64_t, svfloat32x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_f16))) +void svst2_vnum_f16(svbool_t, float16_t *, int64_t, svfloat16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_s32))) +void svst2_vnum_s32(svbool_t, int32_t *, int64_t, svint32x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_s64))) +void svst2_vnum_s64(svbool_t, int64_t *, int64_t, svint64x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_s16))) +void svst2_vnum_s16(svbool_t, int16_t *, int64_t, svint16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_u8))) +void svst3_u8(svbool_t, uint8_t *, svuint8x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_u32))) +void svst3_u32(svbool_t, uint32_t *, svuint32x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_u64))) +void svst3_u64(svbool_t, uint64_t *, svuint64x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_u16))) +void svst3_u16(svbool_t, uint16_t *, svuint16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_s8))) +void svst3_s8(svbool_t, int8_t *, svint8x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_f64))) +void svst3_f64(svbool_t, float64_t *, svfloat64x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_f32))) +void svst3_f32(svbool_t, float32_t *, svfloat32x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_f16))) +void svst3_f16(svbool_t, float16_t *, svfloat16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_s32))) +void svst3_s32(svbool_t, int32_t *, svint32x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_s64))) +void svst3_s64(svbool_t, int64_t *, svint64x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_s16))) +void svst3_s16(svbool_t, int16_t *, svint16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_u8))) +void svst3_vnum_u8(svbool_t, uint8_t *, int64_t, svuint8x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_u32))) +void svst3_vnum_u32(svbool_t, uint32_t *, int64_t, svuint32x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_u64))) +void svst3_vnum_u64(svbool_t, uint64_t *, int64_t, svuint64x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_u16))) +void svst3_vnum_u16(svbool_t, uint16_t *, int64_t, svuint16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_s8))) +void svst3_vnum_s8(svbool_t, int8_t *, int64_t, svint8x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_f64))) +void svst3_vnum_f64(svbool_t, float64_t *, int64_t, svfloat64x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_f32))) +void svst3_vnum_f32(svbool_t, float32_t *, int64_t, svfloat32x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_f16))) +void svst3_vnum_f16(svbool_t, float16_t *, int64_t, svfloat16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_s32))) +void svst3_vnum_s32(svbool_t, int32_t *, int64_t, svint32x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_s64))) +void svst3_vnum_s64(svbool_t, int64_t *, int64_t, svint64x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_s16))) +void svst3_vnum_s16(svbool_t, int16_t *, int64_t, svint16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_u8))) +void svst4_u8(svbool_t, uint8_t *, svuint8x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_u32))) +void svst4_u32(svbool_t, uint32_t *, svuint32x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_u64))) +void svst4_u64(svbool_t, uint64_t *, svuint64x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_u16))) +void svst4_u16(svbool_t, uint16_t *, svuint16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_s8))) +void svst4_s8(svbool_t, int8_t *, svint8x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_f64))) +void svst4_f64(svbool_t, float64_t *, svfloat64x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_f32))) +void svst4_f32(svbool_t, float32_t *, svfloat32x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_f16))) +void svst4_f16(svbool_t, float16_t *, svfloat16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_s32))) +void svst4_s32(svbool_t, int32_t *, svint32x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_s64))) +void svst4_s64(svbool_t, int64_t *, svint64x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_s16))) +void svst4_s16(svbool_t, int16_t *, svint16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_u8))) +void svst4_vnum_u8(svbool_t, uint8_t *, int64_t, svuint8x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_u32))) +void svst4_vnum_u32(svbool_t, uint32_t *, int64_t, svuint32x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_u64))) +void svst4_vnum_u64(svbool_t, uint64_t *, int64_t, svuint64x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_u16))) +void svst4_vnum_u16(svbool_t, uint16_t *, int64_t, svuint16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_s8))) +void svst4_vnum_s8(svbool_t, int8_t *, int64_t, svint8x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_f64))) +void svst4_vnum_f64(svbool_t, float64_t *, int64_t, svfloat64x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_f32))) +void svst4_vnum_f32(svbool_t, float32_t *, int64_t, svfloat32x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_f16))) +void svst4_vnum_f16(svbool_t, float16_t *, int64_t, svfloat16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_s32))) +void svst4_vnum_s32(svbool_t, int32_t *, int64_t, svint32x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_s64))) +void svst4_vnum_s64(svbool_t, int64_t *, int64_t, svint64x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_s16))) +void svst4_vnum_s16(svbool_t, int16_t *, int64_t, svint16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_u8))) +void svstnt1_u8(svbool_t, uint8_t *, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_u32))) +void svstnt1_u32(svbool_t, uint32_t *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_u64))) +void svstnt1_u64(svbool_t, uint64_t *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_u16))) +void svstnt1_u16(svbool_t, uint16_t *, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_s8))) +void svstnt1_s8(svbool_t, int8_t *, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_f64))) +void svstnt1_f64(svbool_t, float64_t *, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_f32))) +void svstnt1_f32(svbool_t, float32_t *, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_f16))) +void svstnt1_f16(svbool_t, float16_t *, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_s32))) +void svstnt1_s32(svbool_t, int32_t *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_s64))) +void svstnt1_s64(svbool_t, int64_t *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_s16))) +void svstnt1_s16(svbool_t, int16_t *, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_u8))) +void svstnt1_vnum_u8(svbool_t, uint8_t *, int64_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_u32))) +void svstnt1_vnum_u32(svbool_t, uint32_t *, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_u64))) +void svstnt1_vnum_u64(svbool_t, uint64_t *, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_u16))) +void svstnt1_vnum_u16(svbool_t, uint16_t *, int64_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_s8))) +void svstnt1_vnum_s8(svbool_t, int8_t *, int64_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_f64))) +void svstnt1_vnum_f64(svbool_t, float64_t *, int64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_f32))) +void svstnt1_vnum_f32(svbool_t, float32_t *, int64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_f16))) +void svstnt1_vnum_f16(svbool_t, float16_t *, int64_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_s32))) +void svstnt1_vnum_s32(svbool_t, int32_t *, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_s64))) +void svstnt1_vnum_s64(svbool_t, int64_t *, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_s16))) +void svstnt1_vnum_s16(svbool_t, int16_t *, int64_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f64_m))) +svfloat64_t svsub_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f32_m))) +svfloat32_t svsub_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f16_m))) +svfloat16_t svsub_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f64_x))) +svfloat64_t svsub_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f32_x))) +svfloat32_t svsub_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f16_x))) +svfloat16_t svsub_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f64_z))) +svfloat64_t svsub_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f32_z))) +svfloat32_t svsub_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f16_z))) +svfloat16_t svsub_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u8_m))) +svuint8_t svsub_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u32_m))) +svuint32_t svsub_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u64_m))) +svuint64_t svsub_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u16_m))) +svuint16_t svsub_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s8_m))) +svint8_t svsub_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s32_m))) +svint32_t svsub_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s64_m))) +svint64_t svsub_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s16_m))) +svint16_t svsub_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u8_x))) +svuint8_t svsub_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u32_x))) +svuint32_t svsub_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u64_x))) +svuint64_t svsub_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u16_x))) +svuint16_t svsub_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s8_x))) +svint8_t svsub_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s32_x))) +svint32_t svsub_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s64_x))) +svint64_t svsub_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s16_x))) +svint16_t svsub_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u8_z))) +svuint8_t svsub_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u32_z))) +svuint32_t svsub_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u64_z))) +svuint64_t svsub_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u16_z))) +svuint16_t svsub_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s8_z))) +svint8_t svsub_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s32_z))) +svint32_t svsub_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s64_z))) +svint64_t svsub_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s16_z))) +svint16_t svsub_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f64_m))) +svfloat64_t svsub_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f32_m))) +svfloat32_t svsub_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f16_m))) +svfloat16_t svsub_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f64_x))) +svfloat64_t svsub_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f32_x))) +svfloat32_t svsub_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f16_x))) +svfloat16_t svsub_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f64_z))) +svfloat64_t svsub_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f32_z))) +svfloat32_t svsub_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f16_z))) +svfloat16_t svsub_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u8_m))) +svuint8_t svsub_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u32_m))) +svuint32_t svsub_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u64_m))) +svuint64_t svsub_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u16_m))) +svuint16_t svsub_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s8_m))) +svint8_t svsub_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s32_m))) +svint32_t svsub_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s64_m))) +svint64_t svsub_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s16_m))) +svint16_t svsub_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u8_x))) +svuint8_t svsub_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u32_x))) +svuint32_t svsub_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u64_x))) +svuint64_t svsub_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u16_x))) +svuint16_t svsub_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s8_x))) +svint8_t svsub_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s32_x))) +svint32_t svsub_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s64_x))) +svint64_t svsub_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s16_x))) +svint16_t svsub_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u8_z))) +svuint8_t svsub_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u32_z))) +svuint32_t svsub_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u64_z))) +svuint64_t svsub_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u16_z))) +svuint16_t svsub_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s8_z))) +svint8_t svsub_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s32_z))) +svint32_t svsub_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s64_z))) +svint64_t svsub_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s16_z))) +svint16_t svsub_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f64_m))) +svfloat64_t svsubr_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f32_m))) +svfloat32_t svsubr_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f16_m))) +svfloat16_t svsubr_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f64_x))) +svfloat64_t svsubr_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f32_x))) +svfloat32_t svsubr_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f16_x))) +svfloat16_t svsubr_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f64_z))) +svfloat64_t svsubr_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f32_z))) +svfloat32_t svsubr_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f16_z))) +svfloat16_t svsubr_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u8_m))) +svuint8_t svsubr_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u32_m))) +svuint32_t svsubr_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u64_m))) +svuint64_t svsubr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u16_m))) +svuint16_t svsubr_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s8_m))) +svint8_t svsubr_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s32_m))) +svint32_t svsubr_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s64_m))) +svint64_t svsubr_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s16_m))) +svint16_t svsubr_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u8_x))) +svuint8_t svsubr_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u32_x))) +svuint32_t svsubr_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u64_x))) +svuint64_t svsubr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u16_x))) +svuint16_t svsubr_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s8_x))) +svint8_t svsubr_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s32_x))) +svint32_t svsubr_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s64_x))) +svint64_t svsubr_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s16_x))) +svint16_t svsubr_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u8_z))) +svuint8_t svsubr_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u32_z))) +svuint32_t svsubr_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u64_z))) +svuint64_t svsubr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u16_z))) +svuint16_t svsubr_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s8_z))) +svint8_t svsubr_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s32_z))) +svint32_t svsubr_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s64_z))) +svint64_t svsubr_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s16_z))) +svint16_t svsubr_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f64_m))) +svfloat64_t svsubr_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f32_m))) +svfloat32_t svsubr_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f16_m))) +svfloat16_t svsubr_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f64_x))) +svfloat64_t svsubr_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f32_x))) +svfloat32_t svsubr_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f16_x))) +svfloat16_t svsubr_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f64_z))) +svfloat64_t svsubr_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f32_z))) +svfloat32_t svsubr_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f16_z))) +svfloat16_t svsubr_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u8_m))) +svuint8_t svsubr_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u32_m))) +svuint32_t svsubr_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u64_m))) +svuint64_t svsubr_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u16_m))) +svuint16_t svsubr_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s8_m))) +svint8_t svsubr_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s32_m))) +svint32_t svsubr_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s64_m))) +svint64_t svsubr_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s16_m))) +svint16_t svsubr_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u8_x))) +svuint8_t svsubr_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u32_x))) +svuint32_t svsubr_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u64_x))) +svuint64_t svsubr_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u16_x))) +svuint16_t svsubr_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s8_x))) +svint8_t svsubr_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s32_x))) +svint32_t svsubr_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s64_x))) +svint64_t svsubr_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s16_x))) +svint16_t svsubr_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u8_z))) +svuint8_t svsubr_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u32_z))) +svuint32_t svsubr_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u64_z))) +svuint64_t svsubr_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u16_z))) +svuint16_t svsubr_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s8_z))) +svint8_t svsubr_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s32_z))) +svint32_t svsubr_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s64_z))) +svint64_t svsubr_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s16_z))) +svint16_t svsubr_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_u8))) +svuint8_t svtbl_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_u32))) +svuint32_t svtbl_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_u64))) +svuint64_t svtbl_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_u16))) +svuint16_t svtbl_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_s8))) +svint8_t svtbl_s8(svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_f64))) +svfloat64_t svtbl_f64(svfloat64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_f32))) +svfloat32_t svtbl_f32(svfloat32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_f16))) +svfloat16_t svtbl_f16(svfloat16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_s32))) +svint32_t svtbl_s32(svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_s64))) +svint64_t svtbl_s64(svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_s16))) +svint16_t svtbl_s16(svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtmad_f64))) +svfloat64_t svtmad_f64(svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtmad_f32))) +svfloat32_t svtmad_f32(svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtmad_f16))) +svfloat16_t svtmad_f16(svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_u8))) +svuint8_t svtrn1_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_u32))) +svuint32_t svtrn1_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_u64))) +svuint64_t svtrn1_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_u16))) +svuint16_t svtrn1_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_s8))) +svint8_t svtrn1_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_f64))) +svfloat64_t svtrn1_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_f32))) +svfloat32_t svtrn1_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_f16))) +svfloat16_t svtrn1_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_s32))) +svint32_t svtrn1_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_s64))) +svint64_t svtrn1_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_s16))) +svint16_t svtrn1_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_b8))) +svbool_t svtrn1_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_b32))) +svbool_t svtrn1_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_b64))) +svbool_t svtrn1_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_b16))) +svbool_t svtrn1_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_u8))) +svuint8_t svtrn2_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_u32))) +svuint32_t svtrn2_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_u64))) +svuint64_t svtrn2_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_u16))) +svuint16_t svtrn2_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_s8))) +svint8_t svtrn2_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_f64))) +svfloat64_t svtrn2_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_f32))) +svfloat32_t svtrn2_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_f16))) +svfloat16_t svtrn2_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_s32))) +svint32_t svtrn2_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_s64))) +svint64_t svtrn2_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_s16))) +svint16_t svtrn2_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_b8))) +svbool_t svtrn2_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_b32))) +svbool_t svtrn2_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_b64))) +svbool_t svtrn2_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_b16))) +svbool_t svtrn2_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtsmul_f64))) +svfloat64_t svtsmul_f64(svfloat64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtsmul_f32))) +svfloat32_t svtsmul_f32(svfloat32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtsmul_f16))) +svfloat16_t svtsmul_f16(svfloat16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtssel_f64))) +svfloat64_t svtssel_f64(svfloat64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtssel_f32))) +svfloat32_t svtssel_f32(svfloat32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtssel_f16))) +svfloat16_t svtssel_f16(svfloat16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_u8))) +svuint8x2_t svundef2_u8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_u32))) +svuint32x2_t svundef2_u32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_u64))) +svuint64x2_t svundef2_u64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_u16))) +svuint16x2_t svundef2_u16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_s8))) +svint8x2_t svundef2_s8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_f64))) +svfloat64x2_t svundef2_f64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_f32))) +svfloat32x2_t svundef2_f32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_f16))) +svfloat16x2_t svundef2_f16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_s32))) +svint32x2_t svundef2_s32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_s64))) +svint64x2_t svundef2_s64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_s16))) +svint16x2_t svundef2_s16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_u8))) +svuint8x3_t svundef3_u8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_u32))) +svuint32x3_t svundef3_u32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_u64))) +svuint64x3_t svundef3_u64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_u16))) +svuint16x3_t svundef3_u16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_s8))) +svint8x3_t svundef3_s8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_f64))) +svfloat64x3_t svundef3_f64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_f32))) +svfloat32x3_t svundef3_f32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_f16))) +svfloat16x3_t svundef3_f16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_s32))) +svint32x3_t svundef3_s32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_s64))) +svint64x3_t svundef3_s64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_s16))) +svint16x3_t svundef3_s16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_u8))) +svuint8x4_t svundef4_u8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_u32))) +svuint32x4_t svundef4_u32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_u64))) +svuint64x4_t svundef4_u64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_u16))) +svuint16x4_t svundef4_u16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_s8))) +svint8x4_t svundef4_s8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_f64))) +svfloat64x4_t svundef4_f64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_f32))) +svfloat32x4_t svundef4_f32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_f16))) +svfloat16x4_t svundef4_f16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_s32))) +svint32x4_t svundef4_s32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_s64))) +svint64x4_t svundef4_s64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_s16))) +svint16x4_t svundef4_s16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_u8))) +svuint8_t svundef_u8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_u32))) +svuint32_t svundef_u32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_u64))) +svuint64_t svundef_u64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_u16))) +svuint16_t svundef_u16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_s8))) +svint8_t svundef_s8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_f64))) +svfloat64_t svundef_f64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_f32))) +svfloat32_t svundef_f32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_f16))) +svfloat16_t svundef_f16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_s32))) +svint32_t svundef_s32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_s64))) +svint64_t svundef_s64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_s16))) +svint16_t svundef_s16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_b))) +svbool_t svunpkhi_b(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_s32))) +svint32_t svunpkhi_s32(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_s64))) +svint64_t svunpkhi_s64(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_s16))) +svint16_t svunpkhi_s16(svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_u32))) +svuint32_t svunpkhi_u32(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_u64))) +svuint64_t svunpkhi_u64(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_u16))) +svuint16_t svunpkhi_u16(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_b))) +svbool_t svunpklo_b(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_s32))) +svint32_t svunpklo_s32(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_s64))) +svint64_t svunpklo_s64(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_s16))) +svint16_t svunpklo_s16(svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_u32))) +svuint32_t svunpklo_u32(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_u64))) +svuint64_t svunpklo_u64(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_u16))) +svuint16_t svunpklo_u16(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_u8))) +svuint8_t svuzp1_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_u32))) +svuint32_t svuzp1_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_u64))) +svuint64_t svuzp1_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_u16))) +svuint16_t svuzp1_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_s8))) +svint8_t svuzp1_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_f64))) +svfloat64_t svuzp1_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_f32))) +svfloat32_t svuzp1_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_f16))) +svfloat16_t svuzp1_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_s32))) +svint32_t svuzp1_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_s64))) +svint64_t svuzp1_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_s16))) +svint16_t svuzp1_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_b8))) +svbool_t svuzp1_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_b32))) +svbool_t svuzp1_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_b64))) +svbool_t svuzp1_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_b16))) +svbool_t svuzp1_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_u8))) +svuint8_t svuzp2_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_u32))) +svuint32_t svuzp2_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_u64))) +svuint64_t svuzp2_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_u16))) +svuint16_t svuzp2_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_s8))) +svint8_t svuzp2_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_f64))) +svfloat64_t svuzp2_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_f32))) +svfloat32_t svuzp2_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_f16))) +svfloat16_t svuzp2_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_s32))) +svint32_t svuzp2_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_s64))) +svint64_t svuzp2_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_s16))) +svint16_t svuzp2_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_b8))) +svbool_t svuzp2_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_b32))) +svbool_t svuzp2_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_b64))) +svbool_t svuzp2_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_b16))) +svbool_t svuzp2_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b8_s32))) +svbool_t svwhilele_b8_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b32_s32))) +svbool_t svwhilele_b32_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b64_s32))) +svbool_t svwhilele_b64_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b16_s32))) +svbool_t svwhilele_b16_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b8_s64))) +svbool_t svwhilele_b8_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b32_s64))) +svbool_t svwhilele_b32_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b64_s64))) +svbool_t svwhilele_b64_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b16_s64))) +svbool_t svwhilele_b16_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b8_u32))) +svbool_t svwhilele_b8_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b32_u32))) +svbool_t svwhilele_b32_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b64_u32))) +svbool_t svwhilele_b64_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b16_u32))) +svbool_t svwhilele_b16_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b8_u64))) +svbool_t svwhilele_b8_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b32_u64))) +svbool_t svwhilele_b32_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b64_u64))) +svbool_t svwhilele_b64_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b16_u64))) +svbool_t svwhilele_b16_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b8_u32))) +svbool_t svwhilelt_b8_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b32_u32))) +svbool_t svwhilelt_b32_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b64_u32))) +svbool_t svwhilelt_b64_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b16_u32))) +svbool_t svwhilelt_b16_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b8_u64))) +svbool_t svwhilelt_b8_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b32_u64))) +svbool_t svwhilelt_b32_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b64_u64))) +svbool_t svwhilelt_b64_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b16_u64))) +svbool_t svwhilelt_b16_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b8_s32))) +svbool_t svwhilelt_b8_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b32_s32))) +svbool_t svwhilelt_b32_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b64_s32))) +svbool_t svwhilelt_b64_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b16_s32))) +svbool_t svwhilelt_b16_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b8_s64))) +svbool_t svwhilelt_b8_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b32_s64))) +svbool_t svwhilelt_b32_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b64_s64))) +svbool_t svwhilelt_b64_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b16_s64))) +svbool_t svwhilelt_b16_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwrffr))) +void svwrffr(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_u8))) +svuint8_t svzip1_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_u32))) +svuint32_t svzip1_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_u64))) +svuint64_t svzip1_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_u16))) +svuint16_t svzip1_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_s8))) +svint8_t svzip1_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_f64))) +svfloat64_t svzip1_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_f32))) +svfloat32_t svzip1_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_f16))) +svfloat16_t svzip1_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_s32))) +svint32_t svzip1_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_s64))) +svint64_t svzip1_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_s16))) +svint16_t svzip1_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_b8))) +svbool_t svzip1_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_b32))) +svbool_t svzip1_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_b64))) +svbool_t svzip1_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_b16))) +svbool_t svzip1_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_u8))) +svuint8_t svzip2_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_u32))) +svuint32_t svzip2_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_u64))) +svuint64_t svzip2_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_u16))) +svuint16_t svzip2_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_s8))) +svint8_t svzip2_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_f64))) +svfloat64_t svzip2_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_f32))) +svfloat32_t svzip2_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_f16))) +svfloat16_t svzip2_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_s32))) +svint32_t svzip2_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_s64))) +svint64_t svzip2_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_s16))) +svint16_t svzip2_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_b8))) +svbool_t svzip2_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_b32))) +svbool_t svzip2_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_b64))) +svbool_t svzip2_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_b16))) +svbool_t svzip2_b16(svbool_t, svbool_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f64_m))) svfloat64_t svabd_m(svbool_t, svfloat64_t, float64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f32_m))) @@ -10009,7 +13830,8 @@ __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_s64))) int64_t svorv(svbool_t, svint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_s16))) int16_t svorv(svbool_t, svint16_t); -#define svpfalse(...) __builtin_sve_svpfalse_b(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpfalse_b))) +svbool_t svpfalse(); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpfirst_b))) svbool_t svpfirst(svbool_t, svbool_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u32base))) @@ -11877,31 +15699,55 @@ svint64_t svzip2(svint64_t, svint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_s16))) svint16_t svzip2(svint16_t, svint16_t); -#if __ARM_FEATURE_SVE2_BITPERM -#define svbdep_n_u8(...) __builtin_sve_svbdep_n_u8(__VA_ARGS__) -#define svbdep_n_u32(...) __builtin_sve_svbdep_n_u32(__VA_ARGS__) -#define svbdep_n_u64(...) __builtin_sve_svbdep_n_u64(__VA_ARGS__) -#define svbdep_n_u16(...) __builtin_sve_svbdep_n_u16(__VA_ARGS__) -#define svbdep_u8(...) __builtin_sve_svbdep_u8(__VA_ARGS__) -#define svbdep_u32(...) __builtin_sve_svbdep_u32(__VA_ARGS__) -#define svbdep_u64(...) __builtin_sve_svbdep_u64(__VA_ARGS__) -#define svbdep_u16(...) __builtin_sve_svbdep_u16(__VA_ARGS__) -#define svbext_n_u8(...) __builtin_sve_svbext_n_u8(__VA_ARGS__) -#define svbext_n_u32(...) __builtin_sve_svbext_n_u32(__VA_ARGS__) -#define svbext_n_u64(...) __builtin_sve_svbext_n_u64(__VA_ARGS__) -#define svbext_n_u16(...) __builtin_sve_svbext_n_u16(__VA_ARGS__) -#define svbext_u8(...) __builtin_sve_svbext_u8(__VA_ARGS__) -#define svbext_u32(...) __builtin_sve_svbext_u32(__VA_ARGS__) -#define svbext_u64(...) __builtin_sve_svbext_u64(__VA_ARGS__) -#define svbext_u16(...) __builtin_sve_svbext_u16(__VA_ARGS__) -#define svbgrp_n_u8(...) __builtin_sve_svbgrp_n_u8(__VA_ARGS__) -#define svbgrp_n_u32(...) __builtin_sve_svbgrp_n_u32(__VA_ARGS__) -#define svbgrp_n_u64(...) __builtin_sve_svbgrp_n_u64(__VA_ARGS__) -#define svbgrp_n_u16(...) __builtin_sve_svbgrp_n_u16(__VA_ARGS__) -#define svbgrp_u8(...) __builtin_sve_svbgrp_u8(__VA_ARGS__) -#define svbgrp_u32(...) __builtin_sve_svbgrp_u32(__VA_ARGS__) -#define svbgrp_u64(...) __builtin_sve_svbgrp_u64(__VA_ARGS__) -#define svbgrp_u16(...) __builtin_sve_svbgrp_u16(__VA_ARGS__) +#if defined (__ARM_FEATURE_SVE2_BITPERM) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_n_u8))) +svuint8_t svbdep_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_n_u32))) +svuint32_t svbdep_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_n_u64))) +svuint64_t svbdep_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_n_u16))) +svuint16_t svbdep_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_u8))) +svuint8_t svbdep_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_u32))) +svuint32_t svbdep_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_u64))) +svuint64_t svbdep_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_u16))) +svuint16_t svbdep_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_n_u8))) +svuint8_t svbext_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_n_u32))) +svuint32_t svbext_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_n_u64))) +svuint64_t svbext_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_n_u16))) +svuint16_t svbext_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_u8))) +svuint8_t svbext_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_u32))) +svuint32_t svbext_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_u64))) +svuint64_t svbext_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_u16))) +svuint16_t svbext_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_n_u8))) +svuint8_t svbgrp_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_n_u32))) +svuint32_t svbgrp_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_n_u64))) +svuint64_t svbgrp_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_n_u16))) +svuint16_t svbgrp_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_u8))) +svuint8_t svbgrp_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_u32))) +svuint32_t svbgrp_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_u64))) +svuint64_t svbgrp_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_u16))) +svuint16_t svbgrp_u16(svuint16_t, svuint16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_n_u8))) svuint8_t svbdep(svuint8_t, uint8_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_n_u32))) @@ -11950,1866 +15796,3723 @@ __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_u64))) svuint64_t svbgrp(svuint64_t, svuint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_u16))) svuint16_t svbgrp(svuint16_t, svuint16_t); -#endif //__ARM_FEATURE_SVE2_BITPERM +#endif //defined (__ARM_FEATURE_SVE2_BITPERM) #if defined(__ARM_FEATURE_SVE2) -#define svaba_n_s8(...) __builtin_sve_svaba_n_s8(__VA_ARGS__) -#define svaba_n_s32(...) __builtin_sve_svaba_n_s32(__VA_ARGS__) -#define svaba_n_s64(...) __builtin_sve_svaba_n_s64(__VA_ARGS__) -#define svaba_n_s16(...) __builtin_sve_svaba_n_s16(__VA_ARGS__) -#define svaba_n_u8(...) __builtin_sve_svaba_n_u8(__VA_ARGS__) -#define svaba_n_u32(...) __builtin_sve_svaba_n_u32(__VA_ARGS__) -#define svaba_n_u64(...) __builtin_sve_svaba_n_u64(__VA_ARGS__) -#define svaba_n_u16(...) __builtin_sve_svaba_n_u16(__VA_ARGS__) -#define svaba_s8(...) __builtin_sve_svaba_s8(__VA_ARGS__) -#define svaba_s32(...) __builtin_sve_svaba_s32(__VA_ARGS__) -#define svaba_s64(...) __builtin_sve_svaba_s64(__VA_ARGS__) -#define svaba_s16(...) __builtin_sve_svaba_s16(__VA_ARGS__) -#define svaba_u8(...) __builtin_sve_svaba_u8(__VA_ARGS__) -#define svaba_u32(...) __builtin_sve_svaba_u32(__VA_ARGS__) -#define svaba_u64(...) __builtin_sve_svaba_u64(__VA_ARGS__) -#define svaba_u16(...) __builtin_sve_svaba_u16(__VA_ARGS__) -#define svabalb_n_s32(...) __builtin_sve_svabalb_n_s32(__VA_ARGS__) -#define svabalb_n_s64(...) __builtin_sve_svabalb_n_s64(__VA_ARGS__) -#define svabalb_n_s16(...) __builtin_sve_svabalb_n_s16(__VA_ARGS__) -#define svabalb_n_u32(...) __builtin_sve_svabalb_n_u32(__VA_ARGS__) -#define svabalb_n_u64(...) __builtin_sve_svabalb_n_u64(__VA_ARGS__) -#define svabalb_n_u16(...) __builtin_sve_svabalb_n_u16(__VA_ARGS__) -#define svabalb_s32(...) __builtin_sve_svabalb_s32(__VA_ARGS__) -#define svabalb_s64(...) __builtin_sve_svabalb_s64(__VA_ARGS__) -#define svabalb_s16(...) __builtin_sve_svabalb_s16(__VA_ARGS__) -#define svabalb_u32(...) __builtin_sve_svabalb_u32(__VA_ARGS__) -#define svabalb_u64(...) __builtin_sve_svabalb_u64(__VA_ARGS__) -#define svabalb_u16(...) __builtin_sve_svabalb_u16(__VA_ARGS__) -#define svabalt_n_s32(...) __builtin_sve_svabalt_n_s32(__VA_ARGS__) -#define svabalt_n_s64(...) __builtin_sve_svabalt_n_s64(__VA_ARGS__) -#define svabalt_n_s16(...) __builtin_sve_svabalt_n_s16(__VA_ARGS__) -#define svabalt_n_u32(...) __builtin_sve_svabalt_n_u32(__VA_ARGS__) -#define svabalt_n_u64(...) __builtin_sve_svabalt_n_u64(__VA_ARGS__) -#define svabalt_n_u16(...) __builtin_sve_svabalt_n_u16(__VA_ARGS__) -#define svabalt_s32(...) __builtin_sve_svabalt_s32(__VA_ARGS__) -#define svabalt_s64(...) __builtin_sve_svabalt_s64(__VA_ARGS__) -#define svabalt_s16(...) __builtin_sve_svabalt_s16(__VA_ARGS__) -#define svabalt_u32(...) __builtin_sve_svabalt_u32(__VA_ARGS__) -#define svabalt_u64(...) __builtin_sve_svabalt_u64(__VA_ARGS__) -#define svabalt_u16(...) __builtin_sve_svabalt_u16(__VA_ARGS__) -#define svabdlb_n_s32(...) __builtin_sve_svabdlb_n_s32(__VA_ARGS__) -#define svabdlb_n_s64(...) __builtin_sve_svabdlb_n_s64(__VA_ARGS__) -#define svabdlb_n_s16(...) __builtin_sve_svabdlb_n_s16(__VA_ARGS__) -#define svabdlb_n_u32(...) __builtin_sve_svabdlb_n_u32(__VA_ARGS__) -#define svabdlb_n_u64(...) __builtin_sve_svabdlb_n_u64(__VA_ARGS__) -#define svabdlb_n_u16(...) __builtin_sve_svabdlb_n_u16(__VA_ARGS__) -#define svabdlb_s32(...) __builtin_sve_svabdlb_s32(__VA_ARGS__) -#define svabdlb_s64(...) __builtin_sve_svabdlb_s64(__VA_ARGS__) -#define svabdlb_s16(...) __builtin_sve_svabdlb_s16(__VA_ARGS__) -#define svabdlb_u32(...) __builtin_sve_svabdlb_u32(__VA_ARGS__) -#define svabdlb_u64(...) __builtin_sve_svabdlb_u64(__VA_ARGS__) -#define svabdlb_u16(...) __builtin_sve_svabdlb_u16(__VA_ARGS__) -#define svabdlt_n_s32(...) __builtin_sve_svabdlt_n_s32(__VA_ARGS__) -#define svabdlt_n_s64(...) __builtin_sve_svabdlt_n_s64(__VA_ARGS__) -#define svabdlt_n_s16(...) __builtin_sve_svabdlt_n_s16(__VA_ARGS__) -#define svabdlt_n_u32(...) __builtin_sve_svabdlt_n_u32(__VA_ARGS__) -#define svabdlt_n_u64(...) __builtin_sve_svabdlt_n_u64(__VA_ARGS__) -#define svabdlt_n_u16(...) __builtin_sve_svabdlt_n_u16(__VA_ARGS__) -#define svabdlt_s32(...) __builtin_sve_svabdlt_s32(__VA_ARGS__) -#define svabdlt_s64(...) __builtin_sve_svabdlt_s64(__VA_ARGS__) -#define svabdlt_s16(...) __builtin_sve_svabdlt_s16(__VA_ARGS__) -#define svabdlt_u32(...) __builtin_sve_svabdlt_u32(__VA_ARGS__) -#define svabdlt_u64(...) __builtin_sve_svabdlt_u64(__VA_ARGS__) -#define svabdlt_u16(...) __builtin_sve_svabdlt_u16(__VA_ARGS__) -#define svadalp_s32_m(...) __builtin_sve_svadalp_s32_m(__VA_ARGS__) -#define svadalp_s64_m(...) __builtin_sve_svadalp_s64_m(__VA_ARGS__) -#define svadalp_s16_m(...) __builtin_sve_svadalp_s16_m(__VA_ARGS__) -#define svadalp_s32_x(...) __builtin_sve_svadalp_s32_x(__VA_ARGS__) -#define svadalp_s64_x(...) __builtin_sve_svadalp_s64_x(__VA_ARGS__) -#define svadalp_s16_x(...) __builtin_sve_svadalp_s16_x(__VA_ARGS__) -#define svadalp_s32_z(...) __builtin_sve_svadalp_s32_z(__VA_ARGS__) -#define svadalp_s64_z(...) __builtin_sve_svadalp_s64_z(__VA_ARGS__) -#define svadalp_s16_z(...) __builtin_sve_svadalp_s16_z(__VA_ARGS__) -#define svadalp_u32_m(...) __builtin_sve_svadalp_u32_m(__VA_ARGS__) -#define svadalp_u64_m(...) __builtin_sve_svadalp_u64_m(__VA_ARGS__) -#define svadalp_u16_m(...) __builtin_sve_svadalp_u16_m(__VA_ARGS__) -#define svadalp_u32_x(...) __builtin_sve_svadalp_u32_x(__VA_ARGS__) -#define svadalp_u64_x(...) __builtin_sve_svadalp_u64_x(__VA_ARGS__) -#define svadalp_u16_x(...) __builtin_sve_svadalp_u16_x(__VA_ARGS__) -#define svadalp_u32_z(...) __builtin_sve_svadalp_u32_z(__VA_ARGS__) -#define svadalp_u64_z(...) __builtin_sve_svadalp_u64_z(__VA_ARGS__) -#define svadalp_u16_z(...) __builtin_sve_svadalp_u16_z(__VA_ARGS__) -#define svadclb_n_u32(...) __builtin_sve_svadclb_n_u32(__VA_ARGS__) -#define svadclb_n_u64(...) __builtin_sve_svadclb_n_u64(__VA_ARGS__) -#define svadclb_u32(...) __builtin_sve_svadclb_u32(__VA_ARGS__) -#define svadclb_u64(...) __builtin_sve_svadclb_u64(__VA_ARGS__) -#define svadclt_n_u32(...) __builtin_sve_svadclt_n_u32(__VA_ARGS__) -#define svadclt_n_u64(...) __builtin_sve_svadclt_n_u64(__VA_ARGS__) -#define svadclt_u32(...) __builtin_sve_svadclt_u32(__VA_ARGS__) -#define svadclt_u64(...) __builtin_sve_svadclt_u64(__VA_ARGS__) -#define svaddhnb_n_u32(...) __builtin_sve_svaddhnb_n_u32(__VA_ARGS__) -#define svaddhnb_n_u64(...) __builtin_sve_svaddhnb_n_u64(__VA_ARGS__) -#define svaddhnb_n_u16(...) __builtin_sve_svaddhnb_n_u16(__VA_ARGS__) -#define svaddhnb_n_s32(...) __builtin_sve_svaddhnb_n_s32(__VA_ARGS__) -#define svaddhnb_n_s64(...) __builtin_sve_svaddhnb_n_s64(__VA_ARGS__) -#define svaddhnb_n_s16(...) __builtin_sve_svaddhnb_n_s16(__VA_ARGS__) -#define svaddhnb_u32(...) __builtin_sve_svaddhnb_u32(__VA_ARGS__) -#define svaddhnb_u64(...) __builtin_sve_svaddhnb_u64(__VA_ARGS__) -#define svaddhnb_u16(...) __builtin_sve_svaddhnb_u16(__VA_ARGS__) -#define svaddhnb_s32(...) __builtin_sve_svaddhnb_s32(__VA_ARGS__) -#define svaddhnb_s64(...) __builtin_sve_svaddhnb_s64(__VA_ARGS__) -#define svaddhnb_s16(...) __builtin_sve_svaddhnb_s16(__VA_ARGS__) -#define svaddhnt_n_u32(...) __builtin_sve_svaddhnt_n_u32(__VA_ARGS__) -#define svaddhnt_n_u64(...) __builtin_sve_svaddhnt_n_u64(__VA_ARGS__) -#define svaddhnt_n_u16(...) __builtin_sve_svaddhnt_n_u16(__VA_ARGS__) -#define svaddhnt_n_s32(...) __builtin_sve_svaddhnt_n_s32(__VA_ARGS__) -#define svaddhnt_n_s64(...) __builtin_sve_svaddhnt_n_s64(__VA_ARGS__) -#define svaddhnt_n_s16(...) __builtin_sve_svaddhnt_n_s16(__VA_ARGS__) -#define svaddhnt_u32(...) __builtin_sve_svaddhnt_u32(__VA_ARGS__) -#define svaddhnt_u64(...) __builtin_sve_svaddhnt_u64(__VA_ARGS__) -#define svaddhnt_u16(...) __builtin_sve_svaddhnt_u16(__VA_ARGS__) -#define svaddhnt_s32(...) __builtin_sve_svaddhnt_s32(__VA_ARGS__) -#define svaddhnt_s64(...) __builtin_sve_svaddhnt_s64(__VA_ARGS__) -#define svaddhnt_s16(...) __builtin_sve_svaddhnt_s16(__VA_ARGS__) -#define svaddlb_n_s32(...) __builtin_sve_svaddlb_n_s32(__VA_ARGS__) -#define svaddlb_n_s64(...) __builtin_sve_svaddlb_n_s64(__VA_ARGS__) -#define svaddlb_n_s16(...) __builtin_sve_svaddlb_n_s16(__VA_ARGS__) -#define svaddlb_n_u32(...) __builtin_sve_svaddlb_n_u32(__VA_ARGS__) -#define svaddlb_n_u64(...) __builtin_sve_svaddlb_n_u64(__VA_ARGS__) -#define svaddlb_n_u16(...) __builtin_sve_svaddlb_n_u16(__VA_ARGS__) -#define svaddlb_s32(...) __builtin_sve_svaddlb_s32(__VA_ARGS__) -#define svaddlb_s64(...) __builtin_sve_svaddlb_s64(__VA_ARGS__) -#define svaddlb_s16(...) __builtin_sve_svaddlb_s16(__VA_ARGS__) -#define svaddlb_u32(...) __builtin_sve_svaddlb_u32(__VA_ARGS__) -#define svaddlb_u64(...) __builtin_sve_svaddlb_u64(__VA_ARGS__) -#define svaddlb_u16(...) __builtin_sve_svaddlb_u16(__VA_ARGS__) -#define svaddlbt_n_s32(...) __builtin_sve_svaddlbt_n_s32(__VA_ARGS__) -#define svaddlbt_n_s64(...) __builtin_sve_svaddlbt_n_s64(__VA_ARGS__) -#define svaddlbt_n_s16(...) __builtin_sve_svaddlbt_n_s16(__VA_ARGS__) -#define svaddlbt_s32(...) __builtin_sve_svaddlbt_s32(__VA_ARGS__) -#define svaddlbt_s64(...) __builtin_sve_svaddlbt_s64(__VA_ARGS__) -#define svaddlbt_s16(...) __builtin_sve_svaddlbt_s16(__VA_ARGS__) -#define svaddlt_n_s32(...) __builtin_sve_svaddlt_n_s32(__VA_ARGS__) -#define svaddlt_n_s64(...) __builtin_sve_svaddlt_n_s64(__VA_ARGS__) -#define svaddlt_n_s16(...) __builtin_sve_svaddlt_n_s16(__VA_ARGS__) -#define svaddlt_n_u32(...) __builtin_sve_svaddlt_n_u32(__VA_ARGS__) -#define svaddlt_n_u64(...) __builtin_sve_svaddlt_n_u64(__VA_ARGS__) -#define svaddlt_n_u16(...) __builtin_sve_svaddlt_n_u16(__VA_ARGS__) -#define svaddlt_s32(...) __builtin_sve_svaddlt_s32(__VA_ARGS__) -#define svaddlt_s64(...) __builtin_sve_svaddlt_s64(__VA_ARGS__) -#define svaddlt_s16(...) __builtin_sve_svaddlt_s16(__VA_ARGS__) -#define svaddlt_u32(...) __builtin_sve_svaddlt_u32(__VA_ARGS__) -#define svaddlt_u64(...) __builtin_sve_svaddlt_u64(__VA_ARGS__) -#define svaddlt_u16(...) __builtin_sve_svaddlt_u16(__VA_ARGS__) -#define svaddp_f64_m(...) __builtin_sve_svaddp_f64_m(__VA_ARGS__) -#define svaddp_f32_m(...) __builtin_sve_svaddp_f32_m(__VA_ARGS__) -#define svaddp_f16_m(...) __builtin_sve_svaddp_f16_m(__VA_ARGS__) -#define svaddp_f64_x(...) __builtin_sve_svaddp_f64_x(__VA_ARGS__) -#define svaddp_f32_x(...) __builtin_sve_svaddp_f32_x(__VA_ARGS__) -#define svaddp_f16_x(...) __builtin_sve_svaddp_f16_x(__VA_ARGS__) -#define svaddp_u8_m(...) __builtin_sve_svaddp_u8_m(__VA_ARGS__) -#define svaddp_u32_m(...) __builtin_sve_svaddp_u32_m(__VA_ARGS__) -#define svaddp_u64_m(...) __builtin_sve_svaddp_u64_m(__VA_ARGS__) -#define svaddp_u16_m(...) __builtin_sve_svaddp_u16_m(__VA_ARGS__) -#define svaddp_s8_m(...) __builtin_sve_svaddp_s8_m(__VA_ARGS__) -#define svaddp_s32_m(...) __builtin_sve_svaddp_s32_m(__VA_ARGS__) -#define svaddp_s64_m(...) __builtin_sve_svaddp_s64_m(__VA_ARGS__) -#define svaddp_s16_m(...) __builtin_sve_svaddp_s16_m(__VA_ARGS__) -#define svaddp_u8_x(...) __builtin_sve_svaddp_u8_x(__VA_ARGS__) -#define svaddp_u32_x(...) __builtin_sve_svaddp_u32_x(__VA_ARGS__) -#define svaddp_u64_x(...) __builtin_sve_svaddp_u64_x(__VA_ARGS__) -#define svaddp_u16_x(...) __builtin_sve_svaddp_u16_x(__VA_ARGS__) -#define svaddp_s8_x(...) __builtin_sve_svaddp_s8_x(__VA_ARGS__) -#define svaddp_s32_x(...) __builtin_sve_svaddp_s32_x(__VA_ARGS__) -#define svaddp_s64_x(...) __builtin_sve_svaddp_s64_x(__VA_ARGS__) -#define svaddp_s16_x(...) __builtin_sve_svaddp_s16_x(__VA_ARGS__) -#define svaddwb_n_s32(...) __builtin_sve_svaddwb_n_s32(__VA_ARGS__) -#define svaddwb_n_s64(...) __builtin_sve_svaddwb_n_s64(__VA_ARGS__) -#define svaddwb_n_s16(...) __builtin_sve_svaddwb_n_s16(__VA_ARGS__) -#define svaddwb_n_u32(...) __builtin_sve_svaddwb_n_u32(__VA_ARGS__) -#define svaddwb_n_u64(...) __builtin_sve_svaddwb_n_u64(__VA_ARGS__) -#define svaddwb_n_u16(...) __builtin_sve_svaddwb_n_u16(__VA_ARGS__) -#define svaddwb_s32(...) __builtin_sve_svaddwb_s32(__VA_ARGS__) -#define svaddwb_s64(...) __builtin_sve_svaddwb_s64(__VA_ARGS__) -#define svaddwb_s16(...) __builtin_sve_svaddwb_s16(__VA_ARGS__) -#define svaddwb_u32(...) __builtin_sve_svaddwb_u32(__VA_ARGS__) -#define svaddwb_u64(...) __builtin_sve_svaddwb_u64(__VA_ARGS__) -#define svaddwb_u16(...) __builtin_sve_svaddwb_u16(__VA_ARGS__) -#define svaddwt_n_s32(...) __builtin_sve_svaddwt_n_s32(__VA_ARGS__) -#define svaddwt_n_s64(...) __builtin_sve_svaddwt_n_s64(__VA_ARGS__) -#define svaddwt_n_s16(...) __builtin_sve_svaddwt_n_s16(__VA_ARGS__) -#define svaddwt_n_u32(...) __builtin_sve_svaddwt_n_u32(__VA_ARGS__) -#define svaddwt_n_u64(...) __builtin_sve_svaddwt_n_u64(__VA_ARGS__) -#define svaddwt_n_u16(...) __builtin_sve_svaddwt_n_u16(__VA_ARGS__) -#define svaddwt_s32(...) __builtin_sve_svaddwt_s32(__VA_ARGS__) -#define svaddwt_s64(...) __builtin_sve_svaddwt_s64(__VA_ARGS__) -#define svaddwt_s16(...) __builtin_sve_svaddwt_s16(__VA_ARGS__) -#define svaddwt_u32(...) __builtin_sve_svaddwt_u32(__VA_ARGS__) -#define svaddwt_u64(...) __builtin_sve_svaddwt_u64(__VA_ARGS__) -#define svaddwt_u16(...) __builtin_sve_svaddwt_u16(__VA_ARGS__) -#define svbcax_n_u8(...) __builtin_sve_svbcax_n_u8(__VA_ARGS__) -#define svbcax_n_u32(...) __builtin_sve_svbcax_n_u32(__VA_ARGS__) -#define svbcax_n_u64(...) __builtin_sve_svbcax_n_u64(__VA_ARGS__) -#define svbcax_n_u16(...) __builtin_sve_svbcax_n_u16(__VA_ARGS__) -#define svbcax_n_s8(...) __builtin_sve_svbcax_n_s8(__VA_ARGS__) -#define svbcax_n_s32(...) __builtin_sve_svbcax_n_s32(__VA_ARGS__) -#define svbcax_n_s64(...) __builtin_sve_svbcax_n_s64(__VA_ARGS__) -#define svbcax_n_s16(...) __builtin_sve_svbcax_n_s16(__VA_ARGS__) -#define svbcax_u8(...) __builtin_sve_svbcax_u8(__VA_ARGS__) -#define svbcax_u32(...) __builtin_sve_svbcax_u32(__VA_ARGS__) -#define svbcax_u64(...) __builtin_sve_svbcax_u64(__VA_ARGS__) -#define svbcax_u16(...) __builtin_sve_svbcax_u16(__VA_ARGS__) -#define svbcax_s8(...) __builtin_sve_svbcax_s8(__VA_ARGS__) -#define svbcax_s32(...) __builtin_sve_svbcax_s32(__VA_ARGS__) -#define svbcax_s64(...) __builtin_sve_svbcax_s64(__VA_ARGS__) -#define svbcax_s16(...) __builtin_sve_svbcax_s16(__VA_ARGS__) -#define svbsl1n_n_u8(...) __builtin_sve_svbsl1n_n_u8(__VA_ARGS__) -#define svbsl1n_n_u32(...) __builtin_sve_svbsl1n_n_u32(__VA_ARGS__) -#define svbsl1n_n_u64(...) __builtin_sve_svbsl1n_n_u64(__VA_ARGS__) -#define svbsl1n_n_u16(...) __builtin_sve_svbsl1n_n_u16(__VA_ARGS__) -#define svbsl1n_n_s8(...) __builtin_sve_svbsl1n_n_s8(__VA_ARGS__) -#define svbsl1n_n_s32(...) __builtin_sve_svbsl1n_n_s32(__VA_ARGS__) -#define svbsl1n_n_s64(...) __builtin_sve_svbsl1n_n_s64(__VA_ARGS__) -#define svbsl1n_n_s16(...) __builtin_sve_svbsl1n_n_s16(__VA_ARGS__) -#define svbsl1n_u8(...) __builtin_sve_svbsl1n_u8(__VA_ARGS__) -#define svbsl1n_u32(...) __builtin_sve_svbsl1n_u32(__VA_ARGS__) -#define svbsl1n_u64(...) __builtin_sve_svbsl1n_u64(__VA_ARGS__) -#define svbsl1n_u16(...) __builtin_sve_svbsl1n_u16(__VA_ARGS__) -#define svbsl1n_s8(...) __builtin_sve_svbsl1n_s8(__VA_ARGS__) -#define svbsl1n_s32(...) __builtin_sve_svbsl1n_s32(__VA_ARGS__) -#define svbsl1n_s64(...) __builtin_sve_svbsl1n_s64(__VA_ARGS__) -#define svbsl1n_s16(...) __builtin_sve_svbsl1n_s16(__VA_ARGS__) -#define svbsl2n_n_u8(...) __builtin_sve_svbsl2n_n_u8(__VA_ARGS__) -#define svbsl2n_n_u32(...) __builtin_sve_svbsl2n_n_u32(__VA_ARGS__) -#define svbsl2n_n_u64(...) __builtin_sve_svbsl2n_n_u64(__VA_ARGS__) -#define svbsl2n_n_u16(...) __builtin_sve_svbsl2n_n_u16(__VA_ARGS__) -#define svbsl2n_n_s8(...) __builtin_sve_svbsl2n_n_s8(__VA_ARGS__) -#define svbsl2n_n_s32(...) __builtin_sve_svbsl2n_n_s32(__VA_ARGS__) -#define svbsl2n_n_s64(...) __builtin_sve_svbsl2n_n_s64(__VA_ARGS__) -#define svbsl2n_n_s16(...) __builtin_sve_svbsl2n_n_s16(__VA_ARGS__) -#define svbsl2n_u8(...) __builtin_sve_svbsl2n_u8(__VA_ARGS__) -#define svbsl2n_u32(...) __builtin_sve_svbsl2n_u32(__VA_ARGS__) -#define svbsl2n_u64(...) __builtin_sve_svbsl2n_u64(__VA_ARGS__) -#define svbsl2n_u16(...) __builtin_sve_svbsl2n_u16(__VA_ARGS__) -#define svbsl2n_s8(...) __builtin_sve_svbsl2n_s8(__VA_ARGS__) -#define svbsl2n_s32(...) __builtin_sve_svbsl2n_s32(__VA_ARGS__) -#define svbsl2n_s64(...) __builtin_sve_svbsl2n_s64(__VA_ARGS__) -#define svbsl2n_s16(...) __builtin_sve_svbsl2n_s16(__VA_ARGS__) -#define svbsl_n_u8(...) __builtin_sve_svbsl_n_u8(__VA_ARGS__) -#define svbsl_n_u32(...) __builtin_sve_svbsl_n_u32(__VA_ARGS__) -#define svbsl_n_u64(...) __builtin_sve_svbsl_n_u64(__VA_ARGS__) -#define svbsl_n_u16(...) __builtin_sve_svbsl_n_u16(__VA_ARGS__) -#define svbsl_n_s8(...) __builtin_sve_svbsl_n_s8(__VA_ARGS__) -#define svbsl_n_s32(...) __builtin_sve_svbsl_n_s32(__VA_ARGS__) -#define svbsl_n_s64(...) __builtin_sve_svbsl_n_s64(__VA_ARGS__) -#define svbsl_n_s16(...) __builtin_sve_svbsl_n_s16(__VA_ARGS__) -#define svbsl_u8(...) __builtin_sve_svbsl_u8(__VA_ARGS__) -#define svbsl_u32(...) __builtin_sve_svbsl_u32(__VA_ARGS__) -#define svbsl_u64(...) __builtin_sve_svbsl_u64(__VA_ARGS__) -#define svbsl_u16(...) __builtin_sve_svbsl_u16(__VA_ARGS__) -#define svbsl_s8(...) __builtin_sve_svbsl_s8(__VA_ARGS__) -#define svbsl_s32(...) __builtin_sve_svbsl_s32(__VA_ARGS__) -#define svbsl_s64(...) __builtin_sve_svbsl_s64(__VA_ARGS__) -#define svbsl_s16(...) __builtin_sve_svbsl_s16(__VA_ARGS__) -#define svcadd_u8(...) __builtin_sve_svcadd_u8(__VA_ARGS__) -#define svcadd_u32(...) __builtin_sve_svcadd_u32(__VA_ARGS__) -#define svcadd_u64(...) __builtin_sve_svcadd_u64(__VA_ARGS__) -#define svcadd_u16(...) __builtin_sve_svcadd_u16(__VA_ARGS__) -#define svcadd_s8(...) __builtin_sve_svcadd_s8(__VA_ARGS__) -#define svcadd_s32(...) __builtin_sve_svcadd_s32(__VA_ARGS__) -#define svcadd_s64(...) __builtin_sve_svcadd_s64(__VA_ARGS__) -#define svcadd_s16(...) __builtin_sve_svcadd_s16(__VA_ARGS__) -#define svcdot_s32(...) __builtin_sve_svcdot_s32(__VA_ARGS__) -#define svcdot_s64(...) __builtin_sve_svcdot_s64(__VA_ARGS__) -#define svcdot_lane_s32(...) __builtin_sve_svcdot_lane_s32(__VA_ARGS__) -#define svcdot_lane_s64(...) __builtin_sve_svcdot_lane_s64(__VA_ARGS__) -#define svcmla_u8(...) __builtin_sve_svcmla_u8(__VA_ARGS__) -#define svcmla_u32(...) __builtin_sve_svcmla_u32(__VA_ARGS__) -#define svcmla_u64(...) __builtin_sve_svcmla_u64(__VA_ARGS__) -#define svcmla_u16(...) __builtin_sve_svcmla_u16(__VA_ARGS__) -#define svcmla_s8(...) __builtin_sve_svcmla_s8(__VA_ARGS__) -#define svcmla_s32(...) __builtin_sve_svcmla_s32(__VA_ARGS__) -#define svcmla_s64(...) __builtin_sve_svcmla_s64(__VA_ARGS__) -#define svcmla_s16(...) __builtin_sve_svcmla_s16(__VA_ARGS__) -#define svcmla_lane_u32(...) __builtin_sve_svcmla_lane_u32(__VA_ARGS__) -#define svcmla_lane_u16(...) __builtin_sve_svcmla_lane_u16(__VA_ARGS__) -#define svcmla_lane_s32(...) __builtin_sve_svcmla_lane_s32(__VA_ARGS__) -#define svcmla_lane_s16(...) __builtin_sve_svcmla_lane_s16(__VA_ARGS__) -#define svcvtlt_f32_f16_m(...) __builtin_sve_svcvtlt_f32_f16_m(__VA_ARGS__) -#define svcvtlt_f32_f16_x(...) __builtin_sve_svcvtlt_f32_f16_x(__VA_ARGS__) -#define svcvtlt_f64_f32_m(...) __builtin_sve_svcvtlt_f64_f32_m(__VA_ARGS__) -#define svcvtlt_f64_f32_x(...) __builtin_sve_svcvtlt_f64_f32_x(__VA_ARGS__) -#define svcvtnt_f16_f32_m(...) __builtin_sve_svcvtnt_f16_f32_m(__VA_ARGS__) -#define svcvtnt_f32_f64_m(...) __builtin_sve_svcvtnt_f32_f64_m(__VA_ARGS__) -#define svcvtx_f32_f64_m(...) __builtin_sve_svcvtx_f32_f64_m(__VA_ARGS__) -#define svcvtx_f32_f64_x(...) __builtin_sve_svcvtx_f32_f64_x(__VA_ARGS__) -#define svcvtx_f32_f64_z(...) __builtin_sve_svcvtx_f32_f64_z(__VA_ARGS__) -#define svcvtxnt_f32_f64_m(...) __builtin_sve_svcvtxnt_f32_f64_m(__VA_ARGS__) -#define sveor3_n_u8(...) __builtin_sve_sveor3_n_u8(__VA_ARGS__) -#define sveor3_n_u32(...) __builtin_sve_sveor3_n_u32(__VA_ARGS__) -#define sveor3_n_u64(...) __builtin_sve_sveor3_n_u64(__VA_ARGS__) -#define sveor3_n_u16(...) __builtin_sve_sveor3_n_u16(__VA_ARGS__) -#define sveor3_n_s8(...) __builtin_sve_sveor3_n_s8(__VA_ARGS__) -#define sveor3_n_s32(...) __builtin_sve_sveor3_n_s32(__VA_ARGS__) -#define sveor3_n_s64(...) __builtin_sve_sveor3_n_s64(__VA_ARGS__) -#define sveor3_n_s16(...) __builtin_sve_sveor3_n_s16(__VA_ARGS__) -#define sveor3_u8(...) __builtin_sve_sveor3_u8(__VA_ARGS__) -#define sveor3_u32(...) __builtin_sve_sveor3_u32(__VA_ARGS__) -#define sveor3_u64(...) __builtin_sve_sveor3_u64(__VA_ARGS__) -#define sveor3_u16(...) __builtin_sve_sveor3_u16(__VA_ARGS__) -#define sveor3_s8(...) __builtin_sve_sveor3_s8(__VA_ARGS__) -#define sveor3_s32(...) __builtin_sve_sveor3_s32(__VA_ARGS__) -#define sveor3_s64(...) __builtin_sve_sveor3_s64(__VA_ARGS__) -#define sveor3_s16(...) __builtin_sve_sveor3_s16(__VA_ARGS__) -#define sveorbt_n_u8(...) __builtin_sve_sveorbt_n_u8(__VA_ARGS__) -#define sveorbt_n_u32(...) __builtin_sve_sveorbt_n_u32(__VA_ARGS__) -#define sveorbt_n_u64(...) __builtin_sve_sveorbt_n_u64(__VA_ARGS__) -#define sveorbt_n_u16(...) __builtin_sve_sveorbt_n_u16(__VA_ARGS__) -#define sveorbt_n_s8(...) __builtin_sve_sveorbt_n_s8(__VA_ARGS__) -#define sveorbt_n_s32(...) __builtin_sve_sveorbt_n_s32(__VA_ARGS__) -#define sveorbt_n_s64(...) __builtin_sve_sveorbt_n_s64(__VA_ARGS__) -#define sveorbt_n_s16(...) __builtin_sve_sveorbt_n_s16(__VA_ARGS__) -#define sveorbt_u8(...) __builtin_sve_sveorbt_u8(__VA_ARGS__) -#define sveorbt_u32(...) __builtin_sve_sveorbt_u32(__VA_ARGS__) -#define sveorbt_u64(...) __builtin_sve_sveorbt_u64(__VA_ARGS__) -#define sveorbt_u16(...) __builtin_sve_sveorbt_u16(__VA_ARGS__) -#define sveorbt_s8(...) __builtin_sve_sveorbt_s8(__VA_ARGS__) -#define sveorbt_s32(...) __builtin_sve_sveorbt_s32(__VA_ARGS__) -#define sveorbt_s64(...) __builtin_sve_sveorbt_s64(__VA_ARGS__) -#define sveorbt_s16(...) __builtin_sve_sveorbt_s16(__VA_ARGS__) -#define sveortb_n_u8(...) __builtin_sve_sveortb_n_u8(__VA_ARGS__) -#define sveortb_n_u32(...) __builtin_sve_sveortb_n_u32(__VA_ARGS__) -#define sveortb_n_u64(...) __builtin_sve_sveortb_n_u64(__VA_ARGS__) -#define sveortb_n_u16(...) __builtin_sve_sveortb_n_u16(__VA_ARGS__) -#define sveortb_n_s8(...) __builtin_sve_sveortb_n_s8(__VA_ARGS__) -#define sveortb_n_s32(...) __builtin_sve_sveortb_n_s32(__VA_ARGS__) -#define sveortb_n_s64(...) __builtin_sve_sveortb_n_s64(__VA_ARGS__) -#define sveortb_n_s16(...) __builtin_sve_sveortb_n_s16(__VA_ARGS__) -#define sveortb_u8(...) __builtin_sve_sveortb_u8(__VA_ARGS__) -#define sveortb_u32(...) __builtin_sve_sveortb_u32(__VA_ARGS__) -#define sveortb_u64(...) __builtin_sve_sveortb_u64(__VA_ARGS__) -#define sveortb_u16(...) __builtin_sve_sveortb_u16(__VA_ARGS__) -#define sveortb_s8(...) __builtin_sve_sveortb_s8(__VA_ARGS__) -#define sveortb_s32(...) __builtin_sve_sveortb_s32(__VA_ARGS__) -#define sveortb_s64(...) __builtin_sve_sveortb_s64(__VA_ARGS__) -#define sveortb_s16(...) __builtin_sve_sveortb_s16(__VA_ARGS__) -#define svhadd_n_s8_m(...) __builtin_sve_svhadd_n_s8_m(__VA_ARGS__) -#define svhadd_n_s32_m(...) __builtin_sve_svhadd_n_s32_m(__VA_ARGS__) -#define svhadd_n_s64_m(...) __builtin_sve_svhadd_n_s64_m(__VA_ARGS__) -#define svhadd_n_s16_m(...) __builtin_sve_svhadd_n_s16_m(__VA_ARGS__) -#define svhadd_n_s8_x(...) __builtin_sve_svhadd_n_s8_x(__VA_ARGS__) -#define svhadd_n_s32_x(...) __builtin_sve_svhadd_n_s32_x(__VA_ARGS__) -#define svhadd_n_s64_x(...) __builtin_sve_svhadd_n_s64_x(__VA_ARGS__) -#define svhadd_n_s16_x(...) __builtin_sve_svhadd_n_s16_x(__VA_ARGS__) -#define svhadd_n_s8_z(...) __builtin_sve_svhadd_n_s8_z(__VA_ARGS__) -#define svhadd_n_s32_z(...) __builtin_sve_svhadd_n_s32_z(__VA_ARGS__) -#define svhadd_n_s64_z(...) __builtin_sve_svhadd_n_s64_z(__VA_ARGS__) -#define svhadd_n_s16_z(...) __builtin_sve_svhadd_n_s16_z(__VA_ARGS__) -#define svhadd_n_u8_m(...) __builtin_sve_svhadd_n_u8_m(__VA_ARGS__) -#define svhadd_n_u32_m(...) __builtin_sve_svhadd_n_u32_m(__VA_ARGS__) -#define svhadd_n_u64_m(...) __builtin_sve_svhadd_n_u64_m(__VA_ARGS__) -#define svhadd_n_u16_m(...) __builtin_sve_svhadd_n_u16_m(__VA_ARGS__) -#define svhadd_n_u8_x(...) __builtin_sve_svhadd_n_u8_x(__VA_ARGS__) -#define svhadd_n_u32_x(...) __builtin_sve_svhadd_n_u32_x(__VA_ARGS__) -#define svhadd_n_u64_x(...) __builtin_sve_svhadd_n_u64_x(__VA_ARGS__) -#define svhadd_n_u16_x(...) __builtin_sve_svhadd_n_u16_x(__VA_ARGS__) -#define svhadd_n_u8_z(...) __builtin_sve_svhadd_n_u8_z(__VA_ARGS__) -#define svhadd_n_u32_z(...) __builtin_sve_svhadd_n_u32_z(__VA_ARGS__) -#define svhadd_n_u64_z(...) __builtin_sve_svhadd_n_u64_z(__VA_ARGS__) -#define svhadd_n_u16_z(...) __builtin_sve_svhadd_n_u16_z(__VA_ARGS__) -#define svhadd_s8_m(...) __builtin_sve_svhadd_s8_m(__VA_ARGS__) -#define svhadd_s32_m(...) __builtin_sve_svhadd_s32_m(__VA_ARGS__) -#define svhadd_s64_m(...) __builtin_sve_svhadd_s64_m(__VA_ARGS__) -#define svhadd_s16_m(...) __builtin_sve_svhadd_s16_m(__VA_ARGS__) -#define svhadd_s8_x(...) __builtin_sve_svhadd_s8_x(__VA_ARGS__) -#define svhadd_s32_x(...) __builtin_sve_svhadd_s32_x(__VA_ARGS__) -#define svhadd_s64_x(...) __builtin_sve_svhadd_s64_x(__VA_ARGS__) -#define svhadd_s16_x(...) __builtin_sve_svhadd_s16_x(__VA_ARGS__) -#define svhadd_s8_z(...) __builtin_sve_svhadd_s8_z(__VA_ARGS__) -#define svhadd_s32_z(...) __builtin_sve_svhadd_s32_z(__VA_ARGS__) -#define svhadd_s64_z(...) __builtin_sve_svhadd_s64_z(__VA_ARGS__) -#define svhadd_s16_z(...) __builtin_sve_svhadd_s16_z(__VA_ARGS__) -#define svhadd_u8_m(...) __builtin_sve_svhadd_u8_m(__VA_ARGS__) -#define svhadd_u32_m(...) __builtin_sve_svhadd_u32_m(__VA_ARGS__) -#define svhadd_u64_m(...) __builtin_sve_svhadd_u64_m(__VA_ARGS__) -#define svhadd_u16_m(...) __builtin_sve_svhadd_u16_m(__VA_ARGS__) -#define svhadd_u8_x(...) __builtin_sve_svhadd_u8_x(__VA_ARGS__) -#define svhadd_u32_x(...) __builtin_sve_svhadd_u32_x(__VA_ARGS__) -#define svhadd_u64_x(...) __builtin_sve_svhadd_u64_x(__VA_ARGS__) -#define svhadd_u16_x(...) __builtin_sve_svhadd_u16_x(__VA_ARGS__) -#define svhadd_u8_z(...) __builtin_sve_svhadd_u8_z(__VA_ARGS__) -#define svhadd_u32_z(...) __builtin_sve_svhadd_u32_z(__VA_ARGS__) -#define svhadd_u64_z(...) __builtin_sve_svhadd_u64_z(__VA_ARGS__) -#define svhadd_u16_z(...) __builtin_sve_svhadd_u16_z(__VA_ARGS__) -#define svhistcnt_u32_z(...) __builtin_sve_svhistcnt_u32_z(__VA_ARGS__) -#define svhistcnt_u64_z(...) __builtin_sve_svhistcnt_u64_z(__VA_ARGS__) -#define svhistcnt_s32_z(...) __builtin_sve_svhistcnt_s32_z(__VA_ARGS__) -#define svhistcnt_s64_z(...) __builtin_sve_svhistcnt_s64_z(__VA_ARGS__) -#define svhistseg_u8(...) __builtin_sve_svhistseg_u8(__VA_ARGS__) -#define svhistseg_s8(...) __builtin_sve_svhistseg_s8(__VA_ARGS__) -#define svhsub_n_s8_m(...) __builtin_sve_svhsub_n_s8_m(__VA_ARGS__) -#define svhsub_n_s32_m(...) __builtin_sve_svhsub_n_s32_m(__VA_ARGS__) -#define svhsub_n_s64_m(...) __builtin_sve_svhsub_n_s64_m(__VA_ARGS__) -#define svhsub_n_s16_m(...) __builtin_sve_svhsub_n_s16_m(__VA_ARGS__) -#define svhsub_n_s8_x(...) __builtin_sve_svhsub_n_s8_x(__VA_ARGS__) -#define svhsub_n_s32_x(...) __builtin_sve_svhsub_n_s32_x(__VA_ARGS__) -#define svhsub_n_s64_x(...) __builtin_sve_svhsub_n_s64_x(__VA_ARGS__) -#define svhsub_n_s16_x(...) __builtin_sve_svhsub_n_s16_x(__VA_ARGS__) -#define svhsub_n_s8_z(...) __builtin_sve_svhsub_n_s8_z(__VA_ARGS__) -#define svhsub_n_s32_z(...) __builtin_sve_svhsub_n_s32_z(__VA_ARGS__) -#define svhsub_n_s64_z(...) __builtin_sve_svhsub_n_s64_z(__VA_ARGS__) -#define svhsub_n_s16_z(...) __builtin_sve_svhsub_n_s16_z(__VA_ARGS__) -#define svhsub_n_u8_m(...) __builtin_sve_svhsub_n_u8_m(__VA_ARGS__) -#define svhsub_n_u32_m(...) __builtin_sve_svhsub_n_u32_m(__VA_ARGS__) -#define svhsub_n_u64_m(...) __builtin_sve_svhsub_n_u64_m(__VA_ARGS__) -#define svhsub_n_u16_m(...) __builtin_sve_svhsub_n_u16_m(__VA_ARGS__) -#define svhsub_n_u8_x(...) __builtin_sve_svhsub_n_u8_x(__VA_ARGS__) -#define svhsub_n_u32_x(...) __builtin_sve_svhsub_n_u32_x(__VA_ARGS__) -#define svhsub_n_u64_x(...) __builtin_sve_svhsub_n_u64_x(__VA_ARGS__) -#define svhsub_n_u16_x(...) __builtin_sve_svhsub_n_u16_x(__VA_ARGS__) -#define svhsub_n_u8_z(...) __builtin_sve_svhsub_n_u8_z(__VA_ARGS__) -#define svhsub_n_u32_z(...) __builtin_sve_svhsub_n_u32_z(__VA_ARGS__) -#define svhsub_n_u64_z(...) __builtin_sve_svhsub_n_u64_z(__VA_ARGS__) -#define svhsub_n_u16_z(...) __builtin_sve_svhsub_n_u16_z(__VA_ARGS__) -#define svhsub_s8_m(...) __builtin_sve_svhsub_s8_m(__VA_ARGS__) -#define svhsub_s32_m(...) __builtin_sve_svhsub_s32_m(__VA_ARGS__) -#define svhsub_s64_m(...) __builtin_sve_svhsub_s64_m(__VA_ARGS__) -#define svhsub_s16_m(...) __builtin_sve_svhsub_s16_m(__VA_ARGS__) -#define svhsub_s8_x(...) __builtin_sve_svhsub_s8_x(__VA_ARGS__) -#define svhsub_s32_x(...) __builtin_sve_svhsub_s32_x(__VA_ARGS__) -#define svhsub_s64_x(...) __builtin_sve_svhsub_s64_x(__VA_ARGS__) -#define svhsub_s16_x(...) __builtin_sve_svhsub_s16_x(__VA_ARGS__) -#define svhsub_s8_z(...) __builtin_sve_svhsub_s8_z(__VA_ARGS__) -#define svhsub_s32_z(...) __builtin_sve_svhsub_s32_z(__VA_ARGS__) -#define svhsub_s64_z(...) __builtin_sve_svhsub_s64_z(__VA_ARGS__) -#define svhsub_s16_z(...) __builtin_sve_svhsub_s16_z(__VA_ARGS__) -#define svhsub_u8_m(...) __builtin_sve_svhsub_u8_m(__VA_ARGS__) -#define svhsub_u32_m(...) __builtin_sve_svhsub_u32_m(__VA_ARGS__) -#define svhsub_u64_m(...) __builtin_sve_svhsub_u64_m(__VA_ARGS__) -#define svhsub_u16_m(...) __builtin_sve_svhsub_u16_m(__VA_ARGS__) -#define svhsub_u8_x(...) __builtin_sve_svhsub_u8_x(__VA_ARGS__) -#define svhsub_u32_x(...) __builtin_sve_svhsub_u32_x(__VA_ARGS__) -#define svhsub_u64_x(...) __builtin_sve_svhsub_u64_x(__VA_ARGS__) -#define svhsub_u16_x(...) __builtin_sve_svhsub_u16_x(__VA_ARGS__) -#define svhsub_u8_z(...) __builtin_sve_svhsub_u8_z(__VA_ARGS__) -#define svhsub_u32_z(...) __builtin_sve_svhsub_u32_z(__VA_ARGS__) -#define svhsub_u64_z(...) __builtin_sve_svhsub_u64_z(__VA_ARGS__) -#define svhsub_u16_z(...) __builtin_sve_svhsub_u16_z(__VA_ARGS__) -#define svhsubr_n_s8_m(...) __builtin_sve_svhsubr_n_s8_m(__VA_ARGS__) -#define svhsubr_n_s32_m(...) __builtin_sve_svhsubr_n_s32_m(__VA_ARGS__) -#define svhsubr_n_s64_m(...) __builtin_sve_svhsubr_n_s64_m(__VA_ARGS__) -#define svhsubr_n_s16_m(...) __builtin_sve_svhsubr_n_s16_m(__VA_ARGS__) -#define svhsubr_n_s8_x(...) __builtin_sve_svhsubr_n_s8_x(__VA_ARGS__) -#define svhsubr_n_s32_x(...) __builtin_sve_svhsubr_n_s32_x(__VA_ARGS__) -#define svhsubr_n_s64_x(...) __builtin_sve_svhsubr_n_s64_x(__VA_ARGS__) -#define svhsubr_n_s16_x(...) __builtin_sve_svhsubr_n_s16_x(__VA_ARGS__) -#define svhsubr_n_s8_z(...) __builtin_sve_svhsubr_n_s8_z(__VA_ARGS__) -#define svhsubr_n_s32_z(...) __builtin_sve_svhsubr_n_s32_z(__VA_ARGS__) -#define svhsubr_n_s64_z(...) __builtin_sve_svhsubr_n_s64_z(__VA_ARGS__) -#define svhsubr_n_s16_z(...) __builtin_sve_svhsubr_n_s16_z(__VA_ARGS__) -#define svhsubr_n_u8_m(...) __builtin_sve_svhsubr_n_u8_m(__VA_ARGS__) -#define svhsubr_n_u32_m(...) __builtin_sve_svhsubr_n_u32_m(__VA_ARGS__) -#define svhsubr_n_u64_m(...) __builtin_sve_svhsubr_n_u64_m(__VA_ARGS__) -#define svhsubr_n_u16_m(...) __builtin_sve_svhsubr_n_u16_m(__VA_ARGS__) -#define svhsubr_n_u8_x(...) __builtin_sve_svhsubr_n_u8_x(__VA_ARGS__) -#define svhsubr_n_u32_x(...) __builtin_sve_svhsubr_n_u32_x(__VA_ARGS__) -#define svhsubr_n_u64_x(...) __builtin_sve_svhsubr_n_u64_x(__VA_ARGS__) -#define svhsubr_n_u16_x(...) __builtin_sve_svhsubr_n_u16_x(__VA_ARGS__) -#define svhsubr_n_u8_z(...) __builtin_sve_svhsubr_n_u8_z(__VA_ARGS__) -#define svhsubr_n_u32_z(...) __builtin_sve_svhsubr_n_u32_z(__VA_ARGS__) -#define svhsubr_n_u64_z(...) __builtin_sve_svhsubr_n_u64_z(__VA_ARGS__) -#define svhsubr_n_u16_z(...) __builtin_sve_svhsubr_n_u16_z(__VA_ARGS__) -#define svhsubr_s8_m(...) __builtin_sve_svhsubr_s8_m(__VA_ARGS__) -#define svhsubr_s32_m(...) __builtin_sve_svhsubr_s32_m(__VA_ARGS__) -#define svhsubr_s64_m(...) __builtin_sve_svhsubr_s64_m(__VA_ARGS__) -#define svhsubr_s16_m(...) __builtin_sve_svhsubr_s16_m(__VA_ARGS__) -#define svhsubr_s8_x(...) __builtin_sve_svhsubr_s8_x(__VA_ARGS__) -#define svhsubr_s32_x(...) __builtin_sve_svhsubr_s32_x(__VA_ARGS__) -#define svhsubr_s64_x(...) __builtin_sve_svhsubr_s64_x(__VA_ARGS__) -#define svhsubr_s16_x(...) __builtin_sve_svhsubr_s16_x(__VA_ARGS__) -#define svhsubr_s8_z(...) __builtin_sve_svhsubr_s8_z(__VA_ARGS__) -#define svhsubr_s32_z(...) __builtin_sve_svhsubr_s32_z(__VA_ARGS__) -#define svhsubr_s64_z(...) __builtin_sve_svhsubr_s64_z(__VA_ARGS__) -#define svhsubr_s16_z(...) __builtin_sve_svhsubr_s16_z(__VA_ARGS__) -#define svhsubr_u8_m(...) __builtin_sve_svhsubr_u8_m(__VA_ARGS__) -#define svhsubr_u32_m(...) __builtin_sve_svhsubr_u32_m(__VA_ARGS__) -#define svhsubr_u64_m(...) __builtin_sve_svhsubr_u64_m(__VA_ARGS__) -#define svhsubr_u16_m(...) __builtin_sve_svhsubr_u16_m(__VA_ARGS__) -#define svhsubr_u8_x(...) __builtin_sve_svhsubr_u8_x(__VA_ARGS__) -#define svhsubr_u32_x(...) __builtin_sve_svhsubr_u32_x(__VA_ARGS__) -#define svhsubr_u64_x(...) __builtin_sve_svhsubr_u64_x(__VA_ARGS__) -#define svhsubr_u16_x(...) __builtin_sve_svhsubr_u16_x(__VA_ARGS__) -#define svhsubr_u8_z(...) __builtin_sve_svhsubr_u8_z(__VA_ARGS__) -#define svhsubr_u32_z(...) __builtin_sve_svhsubr_u32_z(__VA_ARGS__) -#define svhsubr_u64_z(...) __builtin_sve_svhsubr_u64_z(__VA_ARGS__) -#define svhsubr_u16_z(...) __builtin_sve_svhsubr_u16_z(__VA_ARGS__) -#define svldnt1_gather_u32base_index_u32(...) __builtin_sve_svldnt1_gather_u32base_index_u32(__VA_ARGS__) -#define svldnt1_gather_u64base_index_u64(...) __builtin_sve_svldnt1_gather_u64base_index_u64(__VA_ARGS__) -#define svldnt1_gather_u64base_index_f64(...) __builtin_sve_svldnt1_gather_u64base_index_f64(__VA_ARGS__) -#define svldnt1_gather_u32base_index_f32(...) __builtin_sve_svldnt1_gather_u32base_index_f32(__VA_ARGS__) -#define svldnt1_gather_u32base_index_s32(...) __builtin_sve_svldnt1_gather_u32base_index_s32(__VA_ARGS__) -#define svldnt1_gather_u64base_index_s64(...) __builtin_sve_svldnt1_gather_u64base_index_s64(__VA_ARGS__) -#define svldnt1_gather_u32base_offset_u32(...) __builtin_sve_svldnt1_gather_u32base_offset_u32(__VA_ARGS__) -#define svldnt1_gather_u64base_offset_u64(...) __builtin_sve_svldnt1_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1_gather_u64base_offset_f64(...) __builtin_sve_svldnt1_gather_u64base_offset_f64(__VA_ARGS__) -#define svldnt1_gather_u32base_offset_f32(...) __builtin_sve_svldnt1_gather_u32base_offset_f32(__VA_ARGS__) -#define svldnt1_gather_u32base_offset_s32(...) __builtin_sve_svldnt1_gather_u32base_offset_s32(__VA_ARGS__) -#define svldnt1_gather_u64base_offset_s64(...) __builtin_sve_svldnt1_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1_gather_u32base_u32(...) __builtin_sve_svldnt1_gather_u32base_u32(__VA_ARGS__) -#define svldnt1_gather_u64base_u64(...) __builtin_sve_svldnt1_gather_u64base_u64(__VA_ARGS__) -#define svldnt1_gather_u64base_f64(...) __builtin_sve_svldnt1_gather_u64base_f64(__VA_ARGS__) -#define svldnt1_gather_u32base_f32(...) __builtin_sve_svldnt1_gather_u32base_f32(__VA_ARGS__) -#define svldnt1_gather_u32base_s32(...) __builtin_sve_svldnt1_gather_u32base_s32(__VA_ARGS__) -#define svldnt1_gather_u64base_s64(...) __builtin_sve_svldnt1_gather_u64base_s64(__VA_ARGS__) -#define svldnt1_gather_s64index_u64(...) __builtin_sve_svldnt1_gather_s64index_u64(__VA_ARGS__) -#define svldnt1_gather_s64index_f64(...) __builtin_sve_svldnt1_gather_s64index_f64(__VA_ARGS__) -#define svldnt1_gather_s64index_s64(...) __builtin_sve_svldnt1_gather_s64index_s64(__VA_ARGS__) -#define svldnt1_gather_u64index_u64(...) __builtin_sve_svldnt1_gather_u64index_u64(__VA_ARGS__) -#define svldnt1_gather_u64index_f64(...) __builtin_sve_svldnt1_gather_u64index_f64(__VA_ARGS__) -#define svldnt1_gather_u64index_s64(...) __builtin_sve_svldnt1_gather_u64index_s64(__VA_ARGS__) -#define svldnt1_gather_u32offset_u32(...) __builtin_sve_svldnt1_gather_u32offset_u32(__VA_ARGS__) -#define svldnt1_gather_u32offset_f32(...) __builtin_sve_svldnt1_gather_u32offset_f32(__VA_ARGS__) -#define svldnt1_gather_u32offset_s32(...) __builtin_sve_svldnt1_gather_u32offset_s32(__VA_ARGS__) -#define svldnt1_gather_s64offset_u64(...) __builtin_sve_svldnt1_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1_gather_s64offset_f64(...) __builtin_sve_svldnt1_gather_s64offset_f64(__VA_ARGS__) -#define svldnt1_gather_s64offset_s64(...) __builtin_sve_svldnt1_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1_gather_u64offset_u64(...) __builtin_sve_svldnt1_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1_gather_u64offset_f64(...) __builtin_sve_svldnt1_gather_u64offset_f64(__VA_ARGS__) -#define svldnt1_gather_u64offset_s64(...) __builtin_sve_svldnt1_gather_u64offset_s64(__VA_ARGS__) -#define svldnt1sb_gather_u32base_offset_u32(...) __builtin_sve_svldnt1sb_gather_u32base_offset_u32(__VA_ARGS__) -#define svldnt1sb_gather_u64base_offset_u64(...) __builtin_sve_svldnt1sb_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1sb_gather_u32base_offset_s32(...) __builtin_sve_svldnt1sb_gather_u32base_offset_s32(__VA_ARGS__) -#define svldnt1sb_gather_u64base_offset_s64(...) __builtin_sve_svldnt1sb_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1sb_gather_u32base_u32(...) __builtin_sve_svldnt1sb_gather_u32base_u32(__VA_ARGS__) -#define svldnt1sb_gather_u64base_u64(...) __builtin_sve_svldnt1sb_gather_u64base_u64(__VA_ARGS__) -#define svldnt1sb_gather_u32base_s32(...) __builtin_sve_svldnt1sb_gather_u32base_s32(__VA_ARGS__) -#define svldnt1sb_gather_u64base_s64(...) __builtin_sve_svldnt1sb_gather_u64base_s64(__VA_ARGS__) -#define svldnt1sb_gather_u32offset_u32(...) __builtin_sve_svldnt1sb_gather_u32offset_u32(__VA_ARGS__) -#define svldnt1sb_gather_u32offset_s32(...) __builtin_sve_svldnt1sb_gather_u32offset_s32(__VA_ARGS__) -#define svldnt1sb_gather_s64offset_u64(...) __builtin_sve_svldnt1sb_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1sb_gather_s64offset_s64(...) __builtin_sve_svldnt1sb_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1sb_gather_u64offset_u64(...) __builtin_sve_svldnt1sb_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1sb_gather_u64offset_s64(...) __builtin_sve_svldnt1sb_gather_u64offset_s64(__VA_ARGS__) -#define svldnt1sh_gather_u32base_index_u32(...) __builtin_sve_svldnt1sh_gather_u32base_index_u32(__VA_ARGS__) -#define svldnt1sh_gather_u64base_index_u64(...) __builtin_sve_svldnt1sh_gather_u64base_index_u64(__VA_ARGS__) -#define svldnt1sh_gather_u32base_index_s32(...) __builtin_sve_svldnt1sh_gather_u32base_index_s32(__VA_ARGS__) -#define svldnt1sh_gather_u64base_index_s64(...) __builtin_sve_svldnt1sh_gather_u64base_index_s64(__VA_ARGS__) -#define svldnt1sh_gather_u32base_offset_u32(...) __builtin_sve_svldnt1sh_gather_u32base_offset_u32(__VA_ARGS__) -#define svldnt1sh_gather_u64base_offset_u64(...) __builtin_sve_svldnt1sh_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1sh_gather_u32base_offset_s32(...) __builtin_sve_svldnt1sh_gather_u32base_offset_s32(__VA_ARGS__) -#define svldnt1sh_gather_u64base_offset_s64(...) __builtin_sve_svldnt1sh_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1sh_gather_u32base_u32(...) __builtin_sve_svldnt1sh_gather_u32base_u32(__VA_ARGS__) -#define svldnt1sh_gather_u64base_u64(...) __builtin_sve_svldnt1sh_gather_u64base_u64(__VA_ARGS__) -#define svldnt1sh_gather_u32base_s32(...) __builtin_sve_svldnt1sh_gather_u32base_s32(__VA_ARGS__) -#define svldnt1sh_gather_u64base_s64(...) __builtin_sve_svldnt1sh_gather_u64base_s64(__VA_ARGS__) -#define svldnt1sh_gather_s64index_u64(...) __builtin_sve_svldnt1sh_gather_s64index_u64(__VA_ARGS__) -#define svldnt1sh_gather_s64index_s64(...) __builtin_sve_svldnt1sh_gather_s64index_s64(__VA_ARGS__) -#define svldnt1sh_gather_u64index_u64(...) __builtin_sve_svldnt1sh_gather_u64index_u64(__VA_ARGS__) -#define svldnt1sh_gather_u64index_s64(...) __builtin_sve_svldnt1sh_gather_u64index_s64(__VA_ARGS__) -#define svldnt1sh_gather_u32offset_u32(...) __builtin_sve_svldnt1sh_gather_u32offset_u32(__VA_ARGS__) -#define svldnt1sh_gather_u32offset_s32(...) __builtin_sve_svldnt1sh_gather_u32offset_s32(__VA_ARGS__) -#define svldnt1sh_gather_s64offset_u64(...) __builtin_sve_svldnt1sh_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1sh_gather_s64offset_s64(...) __builtin_sve_svldnt1sh_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1sh_gather_u64offset_u64(...) __builtin_sve_svldnt1sh_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1sh_gather_u64offset_s64(...) __builtin_sve_svldnt1sh_gather_u64offset_s64(__VA_ARGS__) -#define svldnt1sw_gather_u64base_index_u64(...) __builtin_sve_svldnt1sw_gather_u64base_index_u64(__VA_ARGS__) -#define svldnt1sw_gather_u64base_index_s64(...) __builtin_sve_svldnt1sw_gather_u64base_index_s64(__VA_ARGS__) -#define svldnt1sw_gather_u64base_offset_u64(...) __builtin_sve_svldnt1sw_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1sw_gather_u64base_offset_s64(...) __builtin_sve_svldnt1sw_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1sw_gather_u64base_u64(...) __builtin_sve_svldnt1sw_gather_u64base_u64(__VA_ARGS__) -#define svldnt1sw_gather_u64base_s64(...) __builtin_sve_svldnt1sw_gather_u64base_s64(__VA_ARGS__) -#define svldnt1sw_gather_s64index_u64(...) __builtin_sve_svldnt1sw_gather_s64index_u64(__VA_ARGS__) -#define svldnt1sw_gather_s64index_s64(...) __builtin_sve_svldnt1sw_gather_s64index_s64(__VA_ARGS__) -#define svldnt1sw_gather_u64index_u64(...) __builtin_sve_svldnt1sw_gather_u64index_u64(__VA_ARGS__) -#define svldnt1sw_gather_u64index_s64(...) __builtin_sve_svldnt1sw_gather_u64index_s64(__VA_ARGS__) -#define svldnt1sw_gather_s64offset_u64(...) __builtin_sve_svldnt1sw_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1sw_gather_s64offset_s64(...) __builtin_sve_svldnt1sw_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1sw_gather_u64offset_u64(...) __builtin_sve_svldnt1sw_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1sw_gather_u64offset_s64(...) __builtin_sve_svldnt1sw_gather_u64offset_s64(__VA_ARGS__) -#define svldnt1ub_gather_u32base_offset_u32(...) __builtin_sve_svldnt1ub_gather_u32base_offset_u32(__VA_ARGS__) -#define svldnt1ub_gather_u64base_offset_u64(...) __builtin_sve_svldnt1ub_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1ub_gather_u32base_offset_s32(...) __builtin_sve_svldnt1ub_gather_u32base_offset_s32(__VA_ARGS__) -#define svldnt1ub_gather_u64base_offset_s64(...) __builtin_sve_svldnt1ub_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1ub_gather_u32base_u32(...) __builtin_sve_svldnt1ub_gather_u32base_u32(__VA_ARGS__) -#define svldnt1ub_gather_u64base_u64(...) __builtin_sve_svldnt1ub_gather_u64base_u64(__VA_ARGS__) -#define svldnt1ub_gather_u32base_s32(...) __builtin_sve_svldnt1ub_gather_u32base_s32(__VA_ARGS__) -#define svldnt1ub_gather_u64base_s64(...) __builtin_sve_svldnt1ub_gather_u64base_s64(__VA_ARGS__) -#define svldnt1ub_gather_u32offset_u32(...) __builtin_sve_svldnt1ub_gather_u32offset_u32(__VA_ARGS__) -#define svldnt1ub_gather_u32offset_s32(...) __builtin_sve_svldnt1ub_gather_u32offset_s32(__VA_ARGS__) -#define svldnt1ub_gather_s64offset_u64(...) __builtin_sve_svldnt1ub_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1ub_gather_s64offset_s64(...) __builtin_sve_svldnt1ub_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1ub_gather_u64offset_u64(...) __builtin_sve_svldnt1ub_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1ub_gather_u64offset_s64(...) __builtin_sve_svldnt1ub_gather_u64offset_s64(__VA_ARGS__) -#define svldnt1uh_gather_u32base_index_u32(...) __builtin_sve_svldnt1uh_gather_u32base_index_u32(__VA_ARGS__) -#define svldnt1uh_gather_u64base_index_u64(...) __builtin_sve_svldnt1uh_gather_u64base_index_u64(__VA_ARGS__) -#define svldnt1uh_gather_u32base_index_s32(...) __builtin_sve_svldnt1uh_gather_u32base_index_s32(__VA_ARGS__) -#define svldnt1uh_gather_u64base_index_s64(...) __builtin_sve_svldnt1uh_gather_u64base_index_s64(__VA_ARGS__) -#define svldnt1uh_gather_u32base_offset_u32(...) __builtin_sve_svldnt1uh_gather_u32base_offset_u32(__VA_ARGS__) -#define svldnt1uh_gather_u64base_offset_u64(...) __builtin_sve_svldnt1uh_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1uh_gather_u32base_offset_s32(...) __builtin_sve_svldnt1uh_gather_u32base_offset_s32(__VA_ARGS__) -#define svldnt1uh_gather_u64base_offset_s64(...) __builtin_sve_svldnt1uh_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1uh_gather_u32base_u32(...) __builtin_sve_svldnt1uh_gather_u32base_u32(__VA_ARGS__) -#define svldnt1uh_gather_u64base_u64(...) __builtin_sve_svldnt1uh_gather_u64base_u64(__VA_ARGS__) -#define svldnt1uh_gather_u32base_s32(...) __builtin_sve_svldnt1uh_gather_u32base_s32(__VA_ARGS__) -#define svldnt1uh_gather_u64base_s64(...) __builtin_sve_svldnt1uh_gather_u64base_s64(__VA_ARGS__) -#define svldnt1uh_gather_s64index_u64(...) __builtin_sve_svldnt1uh_gather_s64index_u64(__VA_ARGS__) -#define svldnt1uh_gather_s64index_s64(...) __builtin_sve_svldnt1uh_gather_s64index_s64(__VA_ARGS__) -#define svldnt1uh_gather_u64index_u64(...) __builtin_sve_svldnt1uh_gather_u64index_u64(__VA_ARGS__) -#define svldnt1uh_gather_u64index_s64(...) __builtin_sve_svldnt1uh_gather_u64index_s64(__VA_ARGS__) -#define svldnt1uh_gather_u32offset_u32(...) __builtin_sve_svldnt1uh_gather_u32offset_u32(__VA_ARGS__) -#define svldnt1uh_gather_u32offset_s32(...) __builtin_sve_svldnt1uh_gather_u32offset_s32(__VA_ARGS__) -#define svldnt1uh_gather_s64offset_u64(...) __builtin_sve_svldnt1uh_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1uh_gather_s64offset_s64(...) __builtin_sve_svldnt1uh_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1uh_gather_u64offset_u64(...) __builtin_sve_svldnt1uh_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1uh_gather_u64offset_s64(...) __builtin_sve_svldnt1uh_gather_u64offset_s64(__VA_ARGS__) -#define svldnt1uw_gather_u64base_index_u64(...) __builtin_sve_svldnt1uw_gather_u64base_index_u64(__VA_ARGS__) -#define svldnt1uw_gather_u64base_index_s64(...) __builtin_sve_svldnt1uw_gather_u64base_index_s64(__VA_ARGS__) -#define svldnt1uw_gather_u64base_offset_u64(...) __builtin_sve_svldnt1uw_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1uw_gather_u64base_offset_s64(...) __builtin_sve_svldnt1uw_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1uw_gather_u64base_u64(...) __builtin_sve_svldnt1uw_gather_u64base_u64(__VA_ARGS__) -#define svldnt1uw_gather_u64base_s64(...) __builtin_sve_svldnt1uw_gather_u64base_s64(__VA_ARGS__) -#define svldnt1uw_gather_s64index_u64(...) __builtin_sve_svldnt1uw_gather_s64index_u64(__VA_ARGS__) -#define svldnt1uw_gather_s64index_s64(...) __builtin_sve_svldnt1uw_gather_s64index_s64(__VA_ARGS__) -#define svldnt1uw_gather_u64index_u64(...) __builtin_sve_svldnt1uw_gather_u64index_u64(__VA_ARGS__) -#define svldnt1uw_gather_u64index_s64(...) __builtin_sve_svldnt1uw_gather_u64index_s64(__VA_ARGS__) -#define svldnt1uw_gather_s64offset_u64(...) __builtin_sve_svldnt1uw_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1uw_gather_s64offset_s64(...) __builtin_sve_svldnt1uw_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1uw_gather_u64offset_u64(...) __builtin_sve_svldnt1uw_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1uw_gather_u64offset_s64(...) __builtin_sve_svldnt1uw_gather_u64offset_s64(__VA_ARGS__) -#define svlogb_f64_m(...) __builtin_sve_svlogb_f64_m(__VA_ARGS__) -#define svlogb_f32_m(...) __builtin_sve_svlogb_f32_m(__VA_ARGS__) -#define svlogb_f16_m(...) __builtin_sve_svlogb_f16_m(__VA_ARGS__) -#define svlogb_f64_x(...) __builtin_sve_svlogb_f64_x(__VA_ARGS__) -#define svlogb_f32_x(...) __builtin_sve_svlogb_f32_x(__VA_ARGS__) -#define svlogb_f16_x(...) __builtin_sve_svlogb_f16_x(__VA_ARGS__) -#define svlogb_f64_z(...) __builtin_sve_svlogb_f64_z(__VA_ARGS__) -#define svlogb_f32_z(...) __builtin_sve_svlogb_f32_z(__VA_ARGS__) -#define svlogb_f16_z(...) __builtin_sve_svlogb_f16_z(__VA_ARGS__) -#define svmatch_u8(...) __builtin_sve_svmatch_u8(__VA_ARGS__) -#define svmatch_u16(...) __builtin_sve_svmatch_u16(__VA_ARGS__) -#define svmatch_s8(...) __builtin_sve_svmatch_s8(__VA_ARGS__) -#define svmatch_s16(...) __builtin_sve_svmatch_s16(__VA_ARGS__) -#define svmaxnmp_f64_m(...) __builtin_sve_svmaxnmp_f64_m(__VA_ARGS__) -#define svmaxnmp_f32_m(...) __builtin_sve_svmaxnmp_f32_m(__VA_ARGS__) -#define svmaxnmp_f16_m(...) __builtin_sve_svmaxnmp_f16_m(__VA_ARGS__) -#define svmaxnmp_f64_x(...) __builtin_sve_svmaxnmp_f64_x(__VA_ARGS__) -#define svmaxnmp_f32_x(...) __builtin_sve_svmaxnmp_f32_x(__VA_ARGS__) -#define svmaxnmp_f16_x(...) __builtin_sve_svmaxnmp_f16_x(__VA_ARGS__) -#define svmaxp_f64_m(...) __builtin_sve_svmaxp_f64_m(__VA_ARGS__) -#define svmaxp_f32_m(...) __builtin_sve_svmaxp_f32_m(__VA_ARGS__) -#define svmaxp_f16_m(...) __builtin_sve_svmaxp_f16_m(__VA_ARGS__) -#define svmaxp_f64_x(...) __builtin_sve_svmaxp_f64_x(__VA_ARGS__) -#define svmaxp_f32_x(...) __builtin_sve_svmaxp_f32_x(__VA_ARGS__) -#define svmaxp_f16_x(...) __builtin_sve_svmaxp_f16_x(__VA_ARGS__) -#define svmaxp_s8_m(...) __builtin_sve_svmaxp_s8_m(__VA_ARGS__) -#define svmaxp_s32_m(...) __builtin_sve_svmaxp_s32_m(__VA_ARGS__) -#define svmaxp_s64_m(...) __builtin_sve_svmaxp_s64_m(__VA_ARGS__) -#define svmaxp_s16_m(...) __builtin_sve_svmaxp_s16_m(__VA_ARGS__) -#define svmaxp_s8_x(...) __builtin_sve_svmaxp_s8_x(__VA_ARGS__) -#define svmaxp_s32_x(...) __builtin_sve_svmaxp_s32_x(__VA_ARGS__) -#define svmaxp_s64_x(...) __builtin_sve_svmaxp_s64_x(__VA_ARGS__) -#define svmaxp_s16_x(...) __builtin_sve_svmaxp_s16_x(__VA_ARGS__) -#define svmaxp_u8_m(...) __builtin_sve_svmaxp_u8_m(__VA_ARGS__) -#define svmaxp_u32_m(...) __builtin_sve_svmaxp_u32_m(__VA_ARGS__) -#define svmaxp_u64_m(...) __builtin_sve_svmaxp_u64_m(__VA_ARGS__) -#define svmaxp_u16_m(...) __builtin_sve_svmaxp_u16_m(__VA_ARGS__) -#define svmaxp_u8_x(...) __builtin_sve_svmaxp_u8_x(__VA_ARGS__) -#define svmaxp_u32_x(...) __builtin_sve_svmaxp_u32_x(__VA_ARGS__) -#define svmaxp_u64_x(...) __builtin_sve_svmaxp_u64_x(__VA_ARGS__) -#define svmaxp_u16_x(...) __builtin_sve_svmaxp_u16_x(__VA_ARGS__) -#define svminnmp_f64_m(...) __builtin_sve_svminnmp_f64_m(__VA_ARGS__) -#define svminnmp_f32_m(...) __builtin_sve_svminnmp_f32_m(__VA_ARGS__) -#define svminnmp_f16_m(...) __builtin_sve_svminnmp_f16_m(__VA_ARGS__) -#define svminnmp_f64_x(...) __builtin_sve_svminnmp_f64_x(__VA_ARGS__) -#define svminnmp_f32_x(...) __builtin_sve_svminnmp_f32_x(__VA_ARGS__) -#define svminnmp_f16_x(...) __builtin_sve_svminnmp_f16_x(__VA_ARGS__) -#define svminp_f64_m(...) __builtin_sve_svminp_f64_m(__VA_ARGS__) -#define svminp_f32_m(...) __builtin_sve_svminp_f32_m(__VA_ARGS__) -#define svminp_f16_m(...) __builtin_sve_svminp_f16_m(__VA_ARGS__) -#define svminp_f64_x(...) __builtin_sve_svminp_f64_x(__VA_ARGS__) -#define svminp_f32_x(...) __builtin_sve_svminp_f32_x(__VA_ARGS__) -#define svminp_f16_x(...) __builtin_sve_svminp_f16_x(__VA_ARGS__) -#define svminp_s8_m(...) __builtin_sve_svminp_s8_m(__VA_ARGS__) -#define svminp_s32_m(...) __builtin_sve_svminp_s32_m(__VA_ARGS__) -#define svminp_s64_m(...) __builtin_sve_svminp_s64_m(__VA_ARGS__) -#define svminp_s16_m(...) __builtin_sve_svminp_s16_m(__VA_ARGS__) -#define svminp_s8_x(...) __builtin_sve_svminp_s8_x(__VA_ARGS__) -#define svminp_s32_x(...) __builtin_sve_svminp_s32_x(__VA_ARGS__) -#define svminp_s64_x(...) __builtin_sve_svminp_s64_x(__VA_ARGS__) -#define svminp_s16_x(...) __builtin_sve_svminp_s16_x(__VA_ARGS__) -#define svminp_u8_m(...) __builtin_sve_svminp_u8_m(__VA_ARGS__) -#define svminp_u32_m(...) __builtin_sve_svminp_u32_m(__VA_ARGS__) -#define svminp_u64_m(...) __builtin_sve_svminp_u64_m(__VA_ARGS__) -#define svminp_u16_m(...) __builtin_sve_svminp_u16_m(__VA_ARGS__) -#define svminp_u8_x(...) __builtin_sve_svminp_u8_x(__VA_ARGS__) -#define svminp_u32_x(...) __builtin_sve_svminp_u32_x(__VA_ARGS__) -#define svminp_u64_x(...) __builtin_sve_svminp_u64_x(__VA_ARGS__) -#define svminp_u16_x(...) __builtin_sve_svminp_u16_x(__VA_ARGS__) -#define svmla_lane_u32(...) __builtin_sve_svmla_lane_u32(__VA_ARGS__) -#define svmla_lane_u64(...) __builtin_sve_svmla_lane_u64(__VA_ARGS__) -#define svmla_lane_u16(...) __builtin_sve_svmla_lane_u16(__VA_ARGS__) -#define svmla_lane_s32(...) __builtin_sve_svmla_lane_s32(__VA_ARGS__) -#define svmla_lane_s64(...) __builtin_sve_svmla_lane_s64(__VA_ARGS__) -#define svmla_lane_s16(...) __builtin_sve_svmla_lane_s16(__VA_ARGS__) -#define svmlalb_n_f32(...) __builtin_sve_svmlalb_n_f32(__VA_ARGS__) -#define svmlalb_n_s32(...) __builtin_sve_svmlalb_n_s32(__VA_ARGS__) -#define svmlalb_n_s64(...) __builtin_sve_svmlalb_n_s64(__VA_ARGS__) -#define svmlalb_n_s16(...) __builtin_sve_svmlalb_n_s16(__VA_ARGS__) -#define svmlalb_n_u32(...) __builtin_sve_svmlalb_n_u32(__VA_ARGS__) -#define svmlalb_n_u64(...) __builtin_sve_svmlalb_n_u64(__VA_ARGS__) -#define svmlalb_n_u16(...) __builtin_sve_svmlalb_n_u16(__VA_ARGS__) -#define svmlalb_f32(...) __builtin_sve_svmlalb_f32(__VA_ARGS__) -#define svmlalb_s32(...) __builtin_sve_svmlalb_s32(__VA_ARGS__) -#define svmlalb_s64(...) __builtin_sve_svmlalb_s64(__VA_ARGS__) -#define svmlalb_s16(...) __builtin_sve_svmlalb_s16(__VA_ARGS__) -#define svmlalb_u32(...) __builtin_sve_svmlalb_u32(__VA_ARGS__) -#define svmlalb_u64(...) __builtin_sve_svmlalb_u64(__VA_ARGS__) -#define svmlalb_u16(...) __builtin_sve_svmlalb_u16(__VA_ARGS__) -#define svmlalb_lane_f32(...) __builtin_sve_svmlalb_lane_f32(__VA_ARGS__) -#define svmlalb_lane_s32(...) __builtin_sve_svmlalb_lane_s32(__VA_ARGS__) -#define svmlalb_lane_s64(...) __builtin_sve_svmlalb_lane_s64(__VA_ARGS__) -#define svmlalb_lane_u32(...) __builtin_sve_svmlalb_lane_u32(__VA_ARGS__) -#define svmlalb_lane_u64(...) __builtin_sve_svmlalb_lane_u64(__VA_ARGS__) -#define svmlalt_n_f32(...) __builtin_sve_svmlalt_n_f32(__VA_ARGS__) -#define svmlalt_n_s32(...) __builtin_sve_svmlalt_n_s32(__VA_ARGS__) -#define svmlalt_n_s64(...) __builtin_sve_svmlalt_n_s64(__VA_ARGS__) -#define svmlalt_n_s16(...) __builtin_sve_svmlalt_n_s16(__VA_ARGS__) -#define svmlalt_n_u32(...) __builtin_sve_svmlalt_n_u32(__VA_ARGS__) -#define svmlalt_n_u64(...) __builtin_sve_svmlalt_n_u64(__VA_ARGS__) -#define svmlalt_n_u16(...) __builtin_sve_svmlalt_n_u16(__VA_ARGS__) -#define svmlalt_f32(...) __builtin_sve_svmlalt_f32(__VA_ARGS__) -#define svmlalt_s32(...) __builtin_sve_svmlalt_s32(__VA_ARGS__) -#define svmlalt_s64(...) __builtin_sve_svmlalt_s64(__VA_ARGS__) -#define svmlalt_s16(...) __builtin_sve_svmlalt_s16(__VA_ARGS__) -#define svmlalt_u32(...) __builtin_sve_svmlalt_u32(__VA_ARGS__) -#define svmlalt_u64(...) __builtin_sve_svmlalt_u64(__VA_ARGS__) -#define svmlalt_u16(...) __builtin_sve_svmlalt_u16(__VA_ARGS__) -#define svmlalt_lane_f32(...) __builtin_sve_svmlalt_lane_f32(__VA_ARGS__) -#define svmlalt_lane_s32(...) __builtin_sve_svmlalt_lane_s32(__VA_ARGS__) -#define svmlalt_lane_s64(...) __builtin_sve_svmlalt_lane_s64(__VA_ARGS__) -#define svmlalt_lane_u32(...) __builtin_sve_svmlalt_lane_u32(__VA_ARGS__) -#define svmlalt_lane_u64(...) __builtin_sve_svmlalt_lane_u64(__VA_ARGS__) -#define svmls_lane_u32(...) __builtin_sve_svmls_lane_u32(__VA_ARGS__) -#define svmls_lane_u64(...) __builtin_sve_svmls_lane_u64(__VA_ARGS__) -#define svmls_lane_u16(...) __builtin_sve_svmls_lane_u16(__VA_ARGS__) -#define svmls_lane_s32(...) __builtin_sve_svmls_lane_s32(__VA_ARGS__) -#define svmls_lane_s64(...) __builtin_sve_svmls_lane_s64(__VA_ARGS__) -#define svmls_lane_s16(...) __builtin_sve_svmls_lane_s16(__VA_ARGS__) -#define svmlslb_n_f32(...) __builtin_sve_svmlslb_n_f32(__VA_ARGS__) -#define svmlslb_n_s32(...) __builtin_sve_svmlslb_n_s32(__VA_ARGS__) -#define svmlslb_n_s64(...) __builtin_sve_svmlslb_n_s64(__VA_ARGS__) -#define svmlslb_n_s16(...) __builtin_sve_svmlslb_n_s16(__VA_ARGS__) -#define svmlslb_n_u32(...) __builtin_sve_svmlslb_n_u32(__VA_ARGS__) -#define svmlslb_n_u64(...) __builtin_sve_svmlslb_n_u64(__VA_ARGS__) -#define svmlslb_n_u16(...) __builtin_sve_svmlslb_n_u16(__VA_ARGS__) -#define svmlslb_f32(...) __builtin_sve_svmlslb_f32(__VA_ARGS__) -#define svmlslb_s32(...) __builtin_sve_svmlslb_s32(__VA_ARGS__) -#define svmlslb_s64(...) __builtin_sve_svmlslb_s64(__VA_ARGS__) -#define svmlslb_s16(...) __builtin_sve_svmlslb_s16(__VA_ARGS__) -#define svmlslb_u32(...) __builtin_sve_svmlslb_u32(__VA_ARGS__) -#define svmlslb_u64(...) __builtin_sve_svmlslb_u64(__VA_ARGS__) -#define svmlslb_u16(...) __builtin_sve_svmlslb_u16(__VA_ARGS__) -#define svmlslb_lane_f32(...) __builtin_sve_svmlslb_lane_f32(__VA_ARGS__) -#define svmlslb_lane_s32(...) __builtin_sve_svmlslb_lane_s32(__VA_ARGS__) -#define svmlslb_lane_s64(...) __builtin_sve_svmlslb_lane_s64(__VA_ARGS__) -#define svmlslb_lane_u32(...) __builtin_sve_svmlslb_lane_u32(__VA_ARGS__) -#define svmlslb_lane_u64(...) __builtin_sve_svmlslb_lane_u64(__VA_ARGS__) -#define svmlslt_n_f32(...) __builtin_sve_svmlslt_n_f32(__VA_ARGS__) -#define svmlslt_n_s32(...) __builtin_sve_svmlslt_n_s32(__VA_ARGS__) -#define svmlslt_n_s64(...) __builtin_sve_svmlslt_n_s64(__VA_ARGS__) -#define svmlslt_n_s16(...) __builtin_sve_svmlslt_n_s16(__VA_ARGS__) -#define svmlslt_n_u32(...) __builtin_sve_svmlslt_n_u32(__VA_ARGS__) -#define svmlslt_n_u64(...) __builtin_sve_svmlslt_n_u64(__VA_ARGS__) -#define svmlslt_n_u16(...) __builtin_sve_svmlslt_n_u16(__VA_ARGS__) -#define svmlslt_f32(...) __builtin_sve_svmlslt_f32(__VA_ARGS__) -#define svmlslt_s32(...) __builtin_sve_svmlslt_s32(__VA_ARGS__) -#define svmlslt_s64(...) __builtin_sve_svmlslt_s64(__VA_ARGS__) -#define svmlslt_s16(...) __builtin_sve_svmlslt_s16(__VA_ARGS__) -#define svmlslt_u32(...) __builtin_sve_svmlslt_u32(__VA_ARGS__) -#define svmlslt_u64(...) __builtin_sve_svmlslt_u64(__VA_ARGS__) -#define svmlslt_u16(...) __builtin_sve_svmlslt_u16(__VA_ARGS__) -#define svmlslt_lane_f32(...) __builtin_sve_svmlslt_lane_f32(__VA_ARGS__) -#define svmlslt_lane_s32(...) __builtin_sve_svmlslt_lane_s32(__VA_ARGS__) -#define svmlslt_lane_s64(...) __builtin_sve_svmlslt_lane_s64(__VA_ARGS__) -#define svmlslt_lane_u32(...) __builtin_sve_svmlslt_lane_u32(__VA_ARGS__) -#define svmlslt_lane_u64(...) __builtin_sve_svmlslt_lane_u64(__VA_ARGS__) -#define svmovlb_s32(...) __builtin_sve_svmovlb_s32(__VA_ARGS__) -#define svmovlb_s64(...) __builtin_sve_svmovlb_s64(__VA_ARGS__) -#define svmovlb_s16(...) __builtin_sve_svmovlb_s16(__VA_ARGS__) -#define svmovlb_u32(...) __builtin_sve_svmovlb_u32(__VA_ARGS__) -#define svmovlb_u64(...) __builtin_sve_svmovlb_u64(__VA_ARGS__) -#define svmovlb_u16(...) __builtin_sve_svmovlb_u16(__VA_ARGS__) -#define svmovlt_s32(...) __builtin_sve_svmovlt_s32(__VA_ARGS__) -#define svmovlt_s64(...) __builtin_sve_svmovlt_s64(__VA_ARGS__) -#define svmovlt_s16(...) __builtin_sve_svmovlt_s16(__VA_ARGS__) -#define svmovlt_u32(...) __builtin_sve_svmovlt_u32(__VA_ARGS__) -#define svmovlt_u64(...) __builtin_sve_svmovlt_u64(__VA_ARGS__) -#define svmovlt_u16(...) __builtin_sve_svmovlt_u16(__VA_ARGS__) -#define svmul_lane_u32(...) __builtin_sve_svmul_lane_u32(__VA_ARGS__) -#define svmul_lane_u64(...) __builtin_sve_svmul_lane_u64(__VA_ARGS__) -#define svmul_lane_u16(...) __builtin_sve_svmul_lane_u16(__VA_ARGS__) -#define svmul_lane_s32(...) __builtin_sve_svmul_lane_s32(__VA_ARGS__) -#define svmul_lane_s64(...) __builtin_sve_svmul_lane_s64(__VA_ARGS__) -#define svmul_lane_s16(...) __builtin_sve_svmul_lane_s16(__VA_ARGS__) -#define svmullb_n_s32(...) __builtin_sve_svmullb_n_s32(__VA_ARGS__) -#define svmullb_n_s64(...) __builtin_sve_svmullb_n_s64(__VA_ARGS__) -#define svmullb_n_s16(...) __builtin_sve_svmullb_n_s16(__VA_ARGS__) -#define svmullb_n_u32(...) __builtin_sve_svmullb_n_u32(__VA_ARGS__) -#define svmullb_n_u64(...) __builtin_sve_svmullb_n_u64(__VA_ARGS__) -#define svmullb_n_u16(...) __builtin_sve_svmullb_n_u16(__VA_ARGS__) -#define svmullb_s32(...) __builtin_sve_svmullb_s32(__VA_ARGS__) -#define svmullb_s64(...) __builtin_sve_svmullb_s64(__VA_ARGS__) -#define svmullb_s16(...) __builtin_sve_svmullb_s16(__VA_ARGS__) -#define svmullb_u32(...) __builtin_sve_svmullb_u32(__VA_ARGS__) -#define svmullb_u64(...) __builtin_sve_svmullb_u64(__VA_ARGS__) -#define svmullb_u16(...) __builtin_sve_svmullb_u16(__VA_ARGS__) -#define svmullb_lane_s32(...) __builtin_sve_svmullb_lane_s32(__VA_ARGS__) -#define svmullb_lane_s64(...) __builtin_sve_svmullb_lane_s64(__VA_ARGS__) -#define svmullb_lane_u32(...) __builtin_sve_svmullb_lane_u32(__VA_ARGS__) -#define svmullb_lane_u64(...) __builtin_sve_svmullb_lane_u64(__VA_ARGS__) -#define svmullt_n_s32(...) __builtin_sve_svmullt_n_s32(__VA_ARGS__) -#define svmullt_n_s64(...) __builtin_sve_svmullt_n_s64(__VA_ARGS__) -#define svmullt_n_s16(...) __builtin_sve_svmullt_n_s16(__VA_ARGS__) -#define svmullt_n_u32(...) __builtin_sve_svmullt_n_u32(__VA_ARGS__) -#define svmullt_n_u64(...) __builtin_sve_svmullt_n_u64(__VA_ARGS__) -#define svmullt_n_u16(...) __builtin_sve_svmullt_n_u16(__VA_ARGS__) -#define svmullt_s32(...) __builtin_sve_svmullt_s32(__VA_ARGS__) -#define svmullt_s64(...) __builtin_sve_svmullt_s64(__VA_ARGS__) -#define svmullt_s16(...) __builtin_sve_svmullt_s16(__VA_ARGS__) -#define svmullt_u32(...) __builtin_sve_svmullt_u32(__VA_ARGS__) -#define svmullt_u64(...) __builtin_sve_svmullt_u64(__VA_ARGS__) -#define svmullt_u16(...) __builtin_sve_svmullt_u16(__VA_ARGS__) -#define svmullt_lane_s32(...) __builtin_sve_svmullt_lane_s32(__VA_ARGS__) -#define svmullt_lane_s64(...) __builtin_sve_svmullt_lane_s64(__VA_ARGS__) -#define svmullt_lane_u32(...) __builtin_sve_svmullt_lane_u32(__VA_ARGS__) -#define svmullt_lane_u64(...) __builtin_sve_svmullt_lane_u64(__VA_ARGS__) -#define svnbsl_n_u8(...) __builtin_sve_svnbsl_n_u8(__VA_ARGS__) -#define svnbsl_n_u32(...) __builtin_sve_svnbsl_n_u32(__VA_ARGS__) -#define svnbsl_n_u64(...) __builtin_sve_svnbsl_n_u64(__VA_ARGS__) -#define svnbsl_n_u16(...) __builtin_sve_svnbsl_n_u16(__VA_ARGS__) -#define svnbsl_n_s8(...) __builtin_sve_svnbsl_n_s8(__VA_ARGS__) -#define svnbsl_n_s32(...) __builtin_sve_svnbsl_n_s32(__VA_ARGS__) -#define svnbsl_n_s64(...) __builtin_sve_svnbsl_n_s64(__VA_ARGS__) -#define svnbsl_n_s16(...) __builtin_sve_svnbsl_n_s16(__VA_ARGS__) -#define svnbsl_u8(...) __builtin_sve_svnbsl_u8(__VA_ARGS__) -#define svnbsl_u32(...) __builtin_sve_svnbsl_u32(__VA_ARGS__) -#define svnbsl_u64(...) __builtin_sve_svnbsl_u64(__VA_ARGS__) -#define svnbsl_u16(...) __builtin_sve_svnbsl_u16(__VA_ARGS__) -#define svnbsl_s8(...) __builtin_sve_svnbsl_s8(__VA_ARGS__) -#define svnbsl_s32(...) __builtin_sve_svnbsl_s32(__VA_ARGS__) -#define svnbsl_s64(...) __builtin_sve_svnbsl_s64(__VA_ARGS__) -#define svnbsl_s16(...) __builtin_sve_svnbsl_s16(__VA_ARGS__) -#define svnmatch_u8(...) __builtin_sve_svnmatch_u8(__VA_ARGS__) -#define svnmatch_u16(...) __builtin_sve_svnmatch_u16(__VA_ARGS__) -#define svnmatch_s8(...) __builtin_sve_svnmatch_s8(__VA_ARGS__) -#define svnmatch_s16(...) __builtin_sve_svnmatch_s16(__VA_ARGS__) -#define svpmul_n_u8(...) __builtin_sve_svpmul_n_u8(__VA_ARGS__) -#define svpmul_u8(...) __builtin_sve_svpmul_u8(__VA_ARGS__) -#define svpmullb_n_u64(...) __builtin_sve_svpmullb_n_u64(__VA_ARGS__) -#define svpmullb_n_u16(...) __builtin_sve_svpmullb_n_u16(__VA_ARGS__) -#define svpmullb_u64(...) __builtin_sve_svpmullb_u64(__VA_ARGS__) -#define svpmullb_u16(...) __builtin_sve_svpmullb_u16(__VA_ARGS__) -#define svpmullb_pair_n_u8(...) __builtin_sve_svpmullb_pair_n_u8(__VA_ARGS__) -#define svpmullb_pair_n_u32(...) __builtin_sve_svpmullb_pair_n_u32(__VA_ARGS__) -#define svpmullb_pair_u8(...) __builtin_sve_svpmullb_pair_u8(__VA_ARGS__) -#define svpmullb_pair_u32(...) __builtin_sve_svpmullb_pair_u32(__VA_ARGS__) -#define svpmullt_n_u64(...) __builtin_sve_svpmullt_n_u64(__VA_ARGS__) -#define svpmullt_n_u16(...) __builtin_sve_svpmullt_n_u16(__VA_ARGS__) -#define svpmullt_u64(...) __builtin_sve_svpmullt_u64(__VA_ARGS__) -#define svpmullt_u16(...) __builtin_sve_svpmullt_u16(__VA_ARGS__) -#define svpmullt_pair_n_u8(...) __builtin_sve_svpmullt_pair_n_u8(__VA_ARGS__) -#define svpmullt_pair_n_u32(...) __builtin_sve_svpmullt_pair_n_u32(__VA_ARGS__) -#define svpmullt_pair_u8(...) __builtin_sve_svpmullt_pair_u8(__VA_ARGS__) -#define svpmullt_pair_u32(...) __builtin_sve_svpmullt_pair_u32(__VA_ARGS__) -#define svqabs_s8_m(...) __builtin_sve_svqabs_s8_m(__VA_ARGS__) -#define svqabs_s32_m(...) __builtin_sve_svqabs_s32_m(__VA_ARGS__) -#define svqabs_s64_m(...) __builtin_sve_svqabs_s64_m(__VA_ARGS__) -#define svqabs_s16_m(...) __builtin_sve_svqabs_s16_m(__VA_ARGS__) -#define svqabs_s8_x(...) __builtin_sve_svqabs_s8_x(__VA_ARGS__) -#define svqabs_s32_x(...) __builtin_sve_svqabs_s32_x(__VA_ARGS__) -#define svqabs_s64_x(...) __builtin_sve_svqabs_s64_x(__VA_ARGS__) -#define svqabs_s16_x(...) __builtin_sve_svqabs_s16_x(__VA_ARGS__) -#define svqabs_s8_z(...) __builtin_sve_svqabs_s8_z(__VA_ARGS__) -#define svqabs_s32_z(...) __builtin_sve_svqabs_s32_z(__VA_ARGS__) -#define svqabs_s64_z(...) __builtin_sve_svqabs_s64_z(__VA_ARGS__) -#define svqabs_s16_z(...) __builtin_sve_svqabs_s16_z(__VA_ARGS__) -#define svqadd_n_s8_m(...) __builtin_sve_svqadd_n_s8_m(__VA_ARGS__) -#define svqadd_n_s32_m(...) __builtin_sve_svqadd_n_s32_m(__VA_ARGS__) -#define svqadd_n_s64_m(...) __builtin_sve_svqadd_n_s64_m(__VA_ARGS__) -#define svqadd_n_s16_m(...) __builtin_sve_svqadd_n_s16_m(__VA_ARGS__) -#define svqadd_n_s8_x(...) __builtin_sve_svqadd_n_s8_x(__VA_ARGS__) -#define svqadd_n_s32_x(...) __builtin_sve_svqadd_n_s32_x(__VA_ARGS__) -#define svqadd_n_s64_x(...) __builtin_sve_svqadd_n_s64_x(__VA_ARGS__) -#define svqadd_n_s16_x(...) __builtin_sve_svqadd_n_s16_x(__VA_ARGS__) -#define svqadd_n_s8_z(...) __builtin_sve_svqadd_n_s8_z(__VA_ARGS__) -#define svqadd_n_s32_z(...) __builtin_sve_svqadd_n_s32_z(__VA_ARGS__) -#define svqadd_n_s64_z(...) __builtin_sve_svqadd_n_s64_z(__VA_ARGS__) -#define svqadd_n_s16_z(...) __builtin_sve_svqadd_n_s16_z(__VA_ARGS__) -#define svqadd_n_u8_m(...) __builtin_sve_svqadd_n_u8_m(__VA_ARGS__) -#define svqadd_n_u32_m(...) __builtin_sve_svqadd_n_u32_m(__VA_ARGS__) -#define svqadd_n_u64_m(...) __builtin_sve_svqadd_n_u64_m(__VA_ARGS__) -#define svqadd_n_u16_m(...) __builtin_sve_svqadd_n_u16_m(__VA_ARGS__) -#define svqadd_n_u8_x(...) __builtin_sve_svqadd_n_u8_x(__VA_ARGS__) -#define svqadd_n_u32_x(...) __builtin_sve_svqadd_n_u32_x(__VA_ARGS__) -#define svqadd_n_u64_x(...) __builtin_sve_svqadd_n_u64_x(__VA_ARGS__) -#define svqadd_n_u16_x(...) __builtin_sve_svqadd_n_u16_x(__VA_ARGS__) -#define svqadd_n_u8_z(...) __builtin_sve_svqadd_n_u8_z(__VA_ARGS__) -#define svqadd_n_u32_z(...) __builtin_sve_svqadd_n_u32_z(__VA_ARGS__) -#define svqadd_n_u64_z(...) __builtin_sve_svqadd_n_u64_z(__VA_ARGS__) -#define svqadd_n_u16_z(...) __builtin_sve_svqadd_n_u16_z(__VA_ARGS__) -#define svqadd_s8_m(...) __builtin_sve_svqadd_s8_m(__VA_ARGS__) -#define svqadd_s32_m(...) __builtin_sve_svqadd_s32_m(__VA_ARGS__) -#define svqadd_s64_m(...) __builtin_sve_svqadd_s64_m(__VA_ARGS__) -#define svqadd_s16_m(...) __builtin_sve_svqadd_s16_m(__VA_ARGS__) -#define svqadd_s8_x(...) __builtin_sve_svqadd_s8_x(__VA_ARGS__) -#define svqadd_s32_x(...) __builtin_sve_svqadd_s32_x(__VA_ARGS__) -#define svqadd_s64_x(...) __builtin_sve_svqadd_s64_x(__VA_ARGS__) -#define svqadd_s16_x(...) __builtin_sve_svqadd_s16_x(__VA_ARGS__) -#define svqadd_s8_z(...) __builtin_sve_svqadd_s8_z(__VA_ARGS__) -#define svqadd_s32_z(...) __builtin_sve_svqadd_s32_z(__VA_ARGS__) -#define svqadd_s64_z(...) __builtin_sve_svqadd_s64_z(__VA_ARGS__) -#define svqadd_s16_z(...) __builtin_sve_svqadd_s16_z(__VA_ARGS__) -#define svqadd_u8_m(...) __builtin_sve_svqadd_u8_m(__VA_ARGS__) -#define svqadd_u32_m(...) __builtin_sve_svqadd_u32_m(__VA_ARGS__) -#define svqadd_u64_m(...) __builtin_sve_svqadd_u64_m(__VA_ARGS__) -#define svqadd_u16_m(...) __builtin_sve_svqadd_u16_m(__VA_ARGS__) -#define svqadd_u8_x(...) __builtin_sve_svqadd_u8_x(__VA_ARGS__) -#define svqadd_u32_x(...) __builtin_sve_svqadd_u32_x(__VA_ARGS__) -#define svqadd_u64_x(...) __builtin_sve_svqadd_u64_x(__VA_ARGS__) -#define svqadd_u16_x(...) __builtin_sve_svqadd_u16_x(__VA_ARGS__) -#define svqadd_u8_z(...) __builtin_sve_svqadd_u8_z(__VA_ARGS__) -#define svqadd_u32_z(...) __builtin_sve_svqadd_u32_z(__VA_ARGS__) -#define svqadd_u64_z(...) __builtin_sve_svqadd_u64_z(__VA_ARGS__) -#define svqadd_u16_z(...) __builtin_sve_svqadd_u16_z(__VA_ARGS__) -#define svqcadd_s8(...) __builtin_sve_svqcadd_s8(__VA_ARGS__) -#define svqcadd_s32(...) __builtin_sve_svqcadd_s32(__VA_ARGS__) -#define svqcadd_s64(...) __builtin_sve_svqcadd_s64(__VA_ARGS__) -#define svqcadd_s16(...) __builtin_sve_svqcadd_s16(__VA_ARGS__) -#define svqdmlalb_n_s32(...) __builtin_sve_svqdmlalb_n_s32(__VA_ARGS__) -#define svqdmlalb_n_s64(...) __builtin_sve_svqdmlalb_n_s64(__VA_ARGS__) -#define svqdmlalb_n_s16(...) __builtin_sve_svqdmlalb_n_s16(__VA_ARGS__) -#define svqdmlalb_s32(...) __builtin_sve_svqdmlalb_s32(__VA_ARGS__) -#define svqdmlalb_s64(...) __builtin_sve_svqdmlalb_s64(__VA_ARGS__) -#define svqdmlalb_s16(...) __builtin_sve_svqdmlalb_s16(__VA_ARGS__) -#define svqdmlalb_lane_s32(...) __builtin_sve_svqdmlalb_lane_s32(__VA_ARGS__) -#define svqdmlalb_lane_s64(...) __builtin_sve_svqdmlalb_lane_s64(__VA_ARGS__) -#define svqdmlalbt_n_s32(...) __builtin_sve_svqdmlalbt_n_s32(__VA_ARGS__) -#define svqdmlalbt_n_s64(...) __builtin_sve_svqdmlalbt_n_s64(__VA_ARGS__) -#define svqdmlalbt_n_s16(...) __builtin_sve_svqdmlalbt_n_s16(__VA_ARGS__) -#define svqdmlalbt_s32(...) __builtin_sve_svqdmlalbt_s32(__VA_ARGS__) -#define svqdmlalbt_s64(...) __builtin_sve_svqdmlalbt_s64(__VA_ARGS__) -#define svqdmlalbt_s16(...) __builtin_sve_svqdmlalbt_s16(__VA_ARGS__) -#define svqdmlalt_n_s32(...) __builtin_sve_svqdmlalt_n_s32(__VA_ARGS__) -#define svqdmlalt_n_s64(...) __builtin_sve_svqdmlalt_n_s64(__VA_ARGS__) -#define svqdmlalt_n_s16(...) __builtin_sve_svqdmlalt_n_s16(__VA_ARGS__) -#define svqdmlalt_s32(...) __builtin_sve_svqdmlalt_s32(__VA_ARGS__) -#define svqdmlalt_s64(...) __builtin_sve_svqdmlalt_s64(__VA_ARGS__) -#define svqdmlalt_s16(...) __builtin_sve_svqdmlalt_s16(__VA_ARGS__) -#define svqdmlalt_lane_s32(...) __builtin_sve_svqdmlalt_lane_s32(__VA_ARGS__) -#define svqdmlalt_lane_s64(...) __builtin_sve_svqdmlalt_lane_s64(__VA_ARGS__) -#define svqdmlslb_n_s32(...) __builtin_sve_svqdmlslb_n_s32(__VA_ARGS__) -#define svqdmlslb_n_s64(...) __builtin_sve_svqdmlslb_n_s64(__VA_ARGS__) -#define svqdmlslb_n_s16(...) __builtin_sve_svqdmlslb_n_s16(__VA_ARGS__) -#define svqdmlslb_s32(...) __builtin_sve_svqdmlslb_s32(__VA_ARGS__) -#define svqdmlslb_s64(...) __builtin_sve_svqdmlslb_s64(__VA_ARGS__) -#define svqdmlslb_s16(...) __builtin_sve_svqdmlslb_s16(__VA_ARGS__) -#define svqdmlslb_lane_s32(...) __builtin_sve_svqdmlslb_lane_s32(__VA_ARGS__) -#define svqdmlslb_lane_s64(...) __builtin_sve_svqdmlslb_lane_s64(__VA_ARGS__) -#define svqdmlslbt_n_s32(...) __builtin_sve_svqdmlslbt_n_s32(__VA_ARGS__) -#define svqdmlslbt_n_s64(...) __builtin_sve_svqdmlslbt_n_s64(__VA_ARGS__) -#define svqdmlslbt_n_s16(...) __builtin_sve_svqdmlslbt_n_s16(__VA_ARGS__) -#define svqdmlslbt_s32(...) __builtin_sve_svqdmlslbt_s32(__VA_ARGS__) -#define svqdmlslbt_s64(...) __builtin_sve_svqdmlslbt_s64(__VA_ARGS__) -#define svqdmlslbt_s16(...) __builtin_sve_svqdmlslbt_s16(__VA_ARGS__) -#define svqdmlslt_n_s32(...) __builtin_sve_svqdmlslt_n_s32(__VA_ARGS__) -#define svqdmlslt_n_s64(...) __builtin_sve_svqdmlslt_n_s64(__VA_ARGS__) -#define svqdmlslt_n_s16(...) __builtin_sve_svqdmlslt_n_s16(__VA_ARGS__) -#define svqdmlslt_s32(...) __builtin_sve_svqdmlslt_s32(__VA_ARGS__) -#define svqdmlslt_s64(...) __builtin_sve_svqdmlslt_s64(__VA_ARGS__) -#define svqdmlslt_s16(...) __builtin_sve_svqdmlslt_s16(__VA_ARGS__) -#define svqdmlslt_lane_s32(...) __builtin_sve_svqdmlslt_lane_s32(__VA_ARGS__) -#define svqdmlslt_lane_s64(...) __builtin_sve_svqdmlslt_lane_s64(__VA_ARGS__) -#define svqdmulh_n_s8(...) __builtin_sve_svqdmulh_n_s8(__VA_ARGS__) -#define svqdmulh_n_s32(...) __builtin_sve_svqdmulh_n_s32(__VA_ARGS__) -#define svqdmulh_n_s64(...) __builtin_sve_svqdmulh_n_s64(__VA_ARGS__) -#define svqdmulh_n_s16(...) __builtin_sve_svqdmulh_n_s16(__VA_ARGS__) -#define svqdmulh_s8(...) __builtin_sve_svqdmulh_s8(__VA_ARGS__) -#define svqdmulh_s32(...) __builtin_sve_svqdmulh_s32(__VA_ARGS__) -#define svqdmulh_s64(...) __builtin_sve_svqdmulh_s64(__VA_ARGS__) -#define svqdmulh_s16(...) __builtin_sve_svqdmulh_s16(__VA_ARGS__) -#define svqdmulh_lane_s32(...) __builtin_sve_svqdmulh_lane_s32(__VA_ARGS__) -#define svqdmulh_lane_s64(...) __builtin_sve_svqdmulh_lane_s64(__VA_ARGS__) -#define svqdmulh_lane_s16(...) __builtin_sve_svqdmulh_lane_s16(__VA_ARGS__) -#define svqdmullb_n_s32(...) __builtin_sve_svqdmullb_n_s32(__VA_ARGS__) -#define svqdmullb_n_s64(...) __builtin_sve_svqdmullb_n_s64(__VA_ARGS__) -#define svqdmullb_n_s16(...) __builtin_sve_svqdmullb_n_s16(__VA_ARGS__) -#define svqdmullb_s32(...) __builtin_sve_svqdmullb_s32(__VA_ARGS__) -#define svqdmullb_s64(...) __builtin_sve_svqdmullb_s64(__VA_ARGS__) -#define svqdmullb_s16(...) __builtin_sve_svqdmullb_s16(__VA_ARGS__) -#define svqdmullb_lane_s32(...) __builtin_sve_svqdmullb_lane_s32(__VA_ARGS__) -#define svqdmullb_lane_s64(...) __builtin_sve_svqdmullb_lane_s64(__VA_ARGS__) -#define svqdmullt_n_s32(...) __builtin_sve_svqdmullt_n_s32(__VA_ARGS__) -#define svqdmullt_n_s64(...) __builtin_sve_svqdmullt_n_s64(__VA_ARGS__) -#define svqdmullt_n_s16(...) __builtin_sve_svqdmullt_n_s16(__VA_ARGS__) -#define svqdmullt_s32(...) __builtin_sve_svqdmullt_s32(__VA_ARGS__) -#define svqdmullt_s64(...) __builtin_sve_svqdmullt_s64(__VA_ARGS__) -#define svqdmullt_s16(...) __builtin_sve_svqdmullt_s16(__VA_ARGS__) -#define svqdmullt_lane_s32(...) __builtin_sve_svqdmullt_lane_s32(__VA_ARGS__) -#define svqdmullt_lane_s64(...) __builtin_sve_svqdmullt_lane_s64(__VA_ARGS__) -#define svqneg_s8_m(...) __builtin_sve_svqneg_s8_m(__VA_ARGS__) -#define svqneg_s32_m(...) __builtin_sve_svqneg_s32_m(__VA_ARGS__) -#define svqneg_s64_m(...) __builtin_sve_svqneg_s64_m(__VA_ARGS__) -#define svqneg_s16_m(...) __builtin_sve_svqneg_s16_m(__VA_ARGS__) -#define svqneg_s8_x(...) __builtin_sve_svqneg_s8_x(__VA_ARGS__) -#define svqneg_s32_x(...) __builtin_sve_svqneg_s32_x(__VA_ARGS__) -#define svqneg_s64_x(...) __builtin_sve_svqneg_s64_x(__VA_ARGS__) -#define svqneg_s16_x(...) __builtin_sve_svqneg_s16_x(__VA_ARGS__) -#define svqneg_s8_z(...) __builtin_sve_svqneg_s8_z(__VA_ARGS__) -#define svqneg_s32_z(...) __builtin_sve_svqneg_s32_z(__VA_ARGS__) -#define svqneg_s64_z(...) __builtin_sve_svqneg_s64_z(__VA_ARGS__) -#define svqneg_s16_z(...) __builtin_sve_svqneg_s16_z(__VA_ARGS__) -#define svqrdcmlah_s8(...) __builtin_sve_svqrdcmlah_s8(__VA_ARGS__) -#define svqrdcmlah_s32(...) __builtin_sve_svqrdcmlah_s32(__VA_ARGS__) -#define svqrdcmlah_s64(...) __builtin_sve_svqrdcmlah_s64(__VA_ARGS__) -#define svqrdcmlah_s16(...) __builtin_sve_svqrdcmlah_s16(__VA_ARGS__) -#define svqrdcmlah_lane_s32(...) __builtin_sve_svqrdcmlah_lane_s32(__VA_ARGS__) -#define svqrdcmlah_lane_s16(...) __builtin_sve_svqrdcmlah_lane_s16(__VA_ARGS__) -#define svqrdmlah_n_s8(...) __builtin_sve_svqrdmlah_n_s8(__VA_ARGS__) -#define svqrdmlah_n_s32(...) __builtin_sve_svqrdmlah_n_s32(__VA_ARGS__) -#define svqrdmlah_n_s64(...) __builtin_sve_svqrdmlah_n_s64(__VA_ARGS__) -#define svqrdmlah_n_s16(...) __builtin_sve_svqrdmlah_n_s16(__VA_ARGS__) -#define svqrdmlah_s8(...) __builtin_sve_svqrdmlah_s8(__VA_ARGS__) -#define svqrdmlah_s32(...) __builtin_sve_svqrdmlah_s32(__VA_ARGS__) -#define svqrdmlah_s64(...) __builtin_sve_svqrdmlah_s64(__VA_ARGS__) -#define svqrdmlah_s16(...) __builtin_sve_svqrdmlah_s16(__VA_ARGS__) -#define svqrdmlah_lane_s32(...) __builtin_sve_svqrdmlah_lane_s32(__VA_ARGS__) -#define svqrdmlah_lane_s64(...) __builtin_sve_svqrdmlah_lane_s64(__VA_ARGS__) -#define svqrdmlah_lane_s16(...) __builtin_sve_svqrdmlah_lane_s16(__VA_ARGS__) -#define svqrdmlsh_n_s8(...) __builtin_sve_svqrdmlsh_n_s8(__VA_ARGS__) -#define svqrdmlsh_n_s32(...) __builtin_sve_svqrdmlsh_n_s32(__VA_ARGS__) -#define svqrdmlsh_n_s64(...) __builtin_sve_svqrdmlsh_n_s64(__VA_ARGS__) -#define svqrdmlsh_n_s16(...) __builtin_sve_svqrdmlsh_n_s16(__VA_ARGS__) -#define svqrdmlsh_s8(...) __builtin_sve_svqrdmlsh_s8(__VA_ARGS__) -#define svqrdmlsh_s32(...) __builtin_sve_svqrdmlsh_s32(__VA_ARGS__) -#define svqrdmlsh_s64(...) __builtin_sve_svqrdmlsh_s64(__VA_ARGS__) -#define svqrdmlsh_s16(...) __builtin_sve_svqrdmlsh_s16(__VA_ARGS__) -#define svqrdmlsh_lane_s32(...) __builtin_sve_svqrdmlsh_lane_s32(__VA_ARGS__) -#define svqrdmlsh_lane_s64(...) __builtin_sve_svqrdmlsh_lane_s64(__VA_ARGS__) -#define svqrdmlsh_lane_s16(...) __builtin_sve_svqrdmlsh_lane_s16(__VA_ARGS__) -#define svqrdmulh_n_s8(...) __builtin_sve_svqrdmulh_n_s8(__VA_ARGS__) -#define svqrdmulh_n_s32(...) __builtin_sve_svqrdmulh_n_s32(__VA_ARGS__) -#define svqrdmulh_n_s64(...) __builtin_sve_svqrdmulh_n_s64(__VA_ARGS__) -#define svqrdmulh_n_s16(...) __builtin_sve_svqrdmulh_n_s16(__VA_ARGS__) -#define svqrdmulh_s8(...) __builtin_sve_svqrdmulh_s8(__VA_ARGS__) -#define svqrdmulh_s32(...) __builtin_sve_svqrdmulh_s32(__VA_ARGS__) -#define svqrdmulh_s64(...) __builtin_sve_svqrdmulh_s64(__VA_ARGS__) -#define svqrdmulh_s16(...) __builtin_sve_svqrdmulh_s16(__VA_ARGS__) -#define svqrdmulh_lane_s32(...) __builtin_sve_svqrdmulh_lane_s32(__VA_ARGS__) -#define svqrdmulh_lane_s64(...) __builtin_sve_svqrdmulh_lane_s64(__VA_ARGS__) -#define svqrdmulh_lane_s16(...) __builtin_sve_svqrdmulh_lane_s16(__VA_ARGS__) -#define svqrshl_n_s8_m(...) __builtin_sve_svqrshl_n_s8_m(__VA_ARGS__) -#define svqrshl_n_s32_m(...) __builtin_sve_svqrshl_n_s32_m(__VA_ARGS__) -#define svqrshl_n_s64_m(...) __builtin_sve_svqrshl_n_s64_m(__VA_ARGS__) -#define svqrshl_n_s16_m(...) __builtin_sve_svqrshl_n_s16_m(__VA_ARGS__) -#define svqrshl_n_s8_x(...) __builtin_sve_svqrshl_n_s8_x(__VA_ARGS__) -#define svqrshl_n_s32_x(...) __builtin_sve_svqrshl_n_s32_x(__VA_ARGS__) -#define svqrshl_n_s64_x(...) __builtin_sve_svqrshl_n_s64_x(__VA_ARGS__) -#define svqrshl_n_s16_x(...) __builtin_sve_svqrshl_n_s16_x(__VA_ARGS__) -#define svqrshl_n_s8_z(...) __builtin_sve_svqrshl_n_s8_z(__VA_ARGS__) -#define svqrshl_n_s32_z(...) __builtin_sve_svqrshl_n_s32_z(__VA_ARGS__) -#define svqrshl_n_s64_z(...) __builtin_sve_svqrshl_n_s64_z(__VA_ARGS__) -#define svqrshl_n_s16_z(...) __builtin_sve_svqrshl_n_s16_z(__VA_ARGS__) -#define svqrshl_n_u8_m(...) __builtin_sve_svqrshl_n_u8_m(__VA_ARGS__) -#define svqrshl_n_u32_m(...) __builtin_sve_svqrshl_n_u32_m(__VA_ARGS__) -#define svqrshl_n_u64_m(...) __builtin_sve_svqrshl_n_u64_m(__VA_ARGS__) -#define svqrshl_n_u16_m(...) __builtin_sve_svqrshl_n_u16_m(__VA_ARGS__) -#define svqrshl_n_u8_x(...) __builtin_sve_svqrshl_n_u8_x(__VA_ARGS__) -#define svqrshl_n_u32_x(...) __builtin_sve_svqrshl_n_u32_x(__VA_ARGS__) -#define svqrshl_n_u64_x(...) __builtin_sve_svqrshl_n_u64_x(__VA_ARGS__) -#define svqrshl_n_u16_x(...) __builtin_sve_svqrshl_n_u16_x(__VA_ARGS__) -#define svqrshl_n_u8_z(...) __builtin_sve_svqrshl_n_u8_z(__VA_ARGS__) -#define svqrshl_n_u32_z(...) __builtin_sve_svqrshl_n_u32_z(__VA_ARGS__) -#define svqrshl_n_u64_z(...) __builtin_sve_svqrshl_n_u64_z(__VA_ARGS__) -#define svqrshl_n_u16_z(...) __builtin_sve_svqrshl_n_u16_z(__VA_ARGS__) -#define svqrshl_s8_m(...) __builtin_sve_svqrshl_s8_m(__VA_ARGS__) -#define svqrshl_s32_m(...) __builtin_sve_svqrshl_s32_m(__VA_ARGS__) -#define svqrshl_s64_m(...) __builtin_sve_svqrshl_s64_m(__VA_ARGS__) -#define svqrshl_s16_m(...) __builtin_sve_svqrshl_s16_m(__VA_ARGS__) -#define svqrshl_s8_x(...) __builtin_sve_svqrshl_s8_x(__VA_ARGS__) -#define svqrshl_s32_x(...) __builtin_sve_svqrshl_s32_x(__VA_ARGS__) -#define svqrshl_s64_x(...) __builtin_sve_svqrshl_s64_x(__VA_ARGS__) -#define svqrshl_s16_x(...) __builtin_sve_svqrshl_s16_x(__VA_ARGS__) -#define svqrshl_s8_z(...) __builtin_sve_svqrshl_s8_z(__VA_ARGS__) -#define svqrshl_s32_z(...) __builtin_sve_svqrshl_s32_z(__VA_ARGS__) -#define svqrshl_s64_z(...) __builtin_sve_svqrshl_s64_z(__VA_ARGS__) -#define svqrshl_s16_z(...) __builtin_sve_svqrshl_s16_z(__VA_ARGS__) -#define svqrshl_u8_m(...) __builtin_sve_svqrshl_u8_m(__VA_ARGS__) -#define svqrshl_u32_m(...) __builtin_sve_svqrshl_u32_m(__VA_ARGS__) -#define svqrshl_u64_m(...) __builtin_sve_svqrshl_u64_m(__VA_ARGS__) -#define svqrshl_u16_m(...) __builtin_sve_svqrshl_u16_m(__VA_ARGS__) -#define svqrshl_u8_x(...) __builtin_sve_svqrshl_u8_x(__VA_ARGS__) -#define svqrshl_u32_x(...) __builtin_sve_svqrshl_u32_x(__VA_ARGS__) -#define svqrshl_u64_x(...) __builtin_sve_svqrshl_u64_x(__VA_ARGS__) -#define svqrshl_u16_x(...) __builtin_sve_svqrshl_u16_x(__VA_ARGS__) -#define svqrshl_u8_z(...) __builtin_sve_svqrshl_u8_z(__VA_ARGS__) -#define svqrshl_u32_z(...) __builtin_sve_svqrshl_u32_z(__VA_ARGS__) -#define svqrshl_u64_z(...) __builtin_sve_svqrshl_u64_z(__VA_ARGS__) -#define svqrshl_u16_z(...) __builtin_sve_svqrshl_u16_z(__VA_ARGS__) -#define svqrshrnb_n_s32(...) __builtin_sve_svqrshrnb_n_s32(__VA_ARGS__) -#define svqrshrnb_n_s64(...) __builtin_sve_svqrshrnb_n_s64(__VA_ARGS__) -#define svqrshrnb_n_s16(...) __builtin_sve_svqrshrnb_n_s16(__VA_ARGS__) -#define svqrshrnb_n_u32(...) __builtin_sve_svqrshrnb_n_u32(__VA_ARGS__) -#define svqrshrnb_n_u64(...) __builtin_sve_svqrshrnb_n_u64(__VA_ARGS__) -#define svqrshrnb_n_u16(...) __builtin_sve_svqrshrnb_n_u16(__VA_ARGS__) -#define svqrshrnt_n_s32(...) __builtin_sve_svqrshrnt_n_s32(__VA_ARGS__) -#define svqrshrnt_n_s64(...) __builtin_sve_svqrshrnt_n_s64(__VA_ARGS__) -#define svqrshrnt_n_s16(...) __builtin_sve_svqrshrnt_n_s16(__VA_ARGS__) -#define svqrshrnt_n_u32(...) __builtin_sve_svqrshrnt_n_u32(__VA_ARGS__) -#define svqrshrnt_n_u64(...) __builtin_sve_svqrshrnt_n_u64(__VA_ARGS__) -#define svqrshrnt_n_u16(...) __builtin_sve_svqrshrnt_n_u16(__VA_ARGS__) -#define svqrshrunb_n_s32(...) __builtin_sve_svqrshrunb_n_s32(__VA_ARGS__) -#define svqrshrunb_n_s64(...) __builtin_sve_svqrshrunb_n_s64(__VA_ARGS__) -#define svqrshrunb_n_s16(...) __builtin_sve_svqrshrunb_n_s16(__VA_ARGS__) -#define svqrshrunt_n_s32(...) __builtin_sve_svqrshrunt_n_s32(__VA_ARGS__) -#define svqrshrunt_n_s64(...) __builtin_sve_svqrshrunt_n_s64(__VA_ARGS__) -#define svqrshrunt_n_s16(...) __builtin_sve_svqrshrunt_n_s16(__VA_ARGS__) -#define svqshl_n_s8_m(...) __builtin_sve_svqshl_n_s8_m(__VA_ARGS__) -#define svqshl_n_s32_m(...) __builtin_sve_svqshl_n_s32_m(__VA_ARGS__) -#define svqshl_n_s64_m(...) __builtin_sve_svqshl_n_s64_m(__VA_ARGS__) -#define svqshl_n_s16_m(...) __builtin_sve_svqshl_n_s16_m(__VA_ARGS__) -#define svqshl_n_s8_x(...) __builtin_sve_svqshl_n_s8_x(__VA_ARGS__) -#define svqshl_n_s32_x(...) __builtin_sve_svqshl_n_s32_x(__VA_ARGS__) -#define svqshl_n_s64_x(...) __builtin_sve_svqshl_n_s64_x(__VA_ARGS__) -#define svqshl_n_s16_x(...) __builtin_sve_svqshl_n_s16_x(__VA_ARGS__) -#define svqshl_n_s8_z(...) __builtin_sve_svqshl_n_s8_z(__VA_ARGS__) -#define svqshl_n_s32_z(...) __builtin_sve_svqshl_n_s32_z(__VA_ARGS__) -#define svqshl_n_s64_z(...) __builtin_sve_svqshl_n_s64_z(__VA_ARGS__) -#define svqshl_n_s16_z(...) __builtin_sve_svqshl_n_s16_z(__VA_ARGS__) -#define svqshl_n_u8_m(...) __builtin_sve_svqshl_n_u8_m(__VA_ARGS__) -#define svqshl_n_u32_m(...) __builtin_sve_svqshl_n_u32_m(__VA_ARGS__) -#define svqshl_n_u64_m(...) __builtin_sve_svqshl_n_u64_m(__VA_ARGS__) -#define svqshl_n_u16_m(...) __builtin_sve_svqshl_n_u16_m(__VA_ARGS__) -#define svqshl_n_u8_x(...) __builtin_sve_svqshl_n_u8_x(__VA_ARGS__) -#define svqshl_n_u32_x(...) __builtin_sve_svqshl_n_u32_x(__VA_ARGS__) -#define svqshl_n_u64_x(...) __builtin_sve_svqshl_n_u64_x(__VA_ARGS__) -#define svqshl_n_u16_x(...) __builtin_sve_svqshl_n_u16_x(__VA_ARGS__) -#define svqshl_n_u8_z(...) __builtin_sve_svqshl_n_u8_z(__VA_ARGS__) -#define svqshl_n_u32_z(...) __builtin_sve_svqshl_n_u32_z(__VA_ARGS__) -#define svqshl_n_u64_z(...) __builtin_sve_svqshl_n_u64_z(__VA_ARGS__) -#define svqshl_n_u16_z(...) __builtin_sve_svqshl_n_u16_z(__VA_ARGS__) -#define svqshl_s8_m(...) __builtin_sve_svqshl_s8_m(__VA_ARGS__) -#define svqshl_s32_m(...) __builtin_sve_svqshl_s32_m(__VA_ARGS__) -#define svqshl_s64_m(...) __builtin_sve_svqshl_s64_m(__VA_ARGS__) -#define svqshl_s16_m(...) __builtin_sve_svqshl_s16_m(__VA_ARGS__) -#define svqshl_s8_x(...) __builtin_sve_svqshl_s8_x(__VA_ARGS__) -#define svqshl_s32_x(...) __builtin_sve_svqshl_s32_x(__VA_ARGS__) -#define svqshl_s64_x(...) __builtin_sve_svqshl_s64_x(__VA_ARGS__) -#define svqshl_s16_x(...) __builtin_sve_svqshl_s16_x(__VA_ARGS__) -#define svqshl_s8_z(...) __builtin_sve_svqshl_s8_z(__VA_ARGS__) -#define svqshl_s32_z(...) __builtin_sve_svqshl_s32_z(__VA_ARGS__) -#define svqshl_s64_z(...) __builtin_sve_svqshl_s64_z(__VA_ARGS__) -#define svqshl_s16_z(...) __builtin_sve_svqshl_s16_z(__VA_ARGS__) -#define svqshl_u8_m(...) __builtin_sve_svqshl_u8_m(__VA_ARGS__) -#define svqshl_u32_m(...) __builtin_sve_svqshl_u32_m(__VA_ARGS__) -#define svqshl_u64_m(...) __builtin_sve_svqshl_u64_m(__VA_ARGS__) -#define svqshl_u16_m(...) __builtin_sve_svqshl_u16_m(__VA_ARGS__) -#define svqshl_u8_x(...) __builtin_sve_svqshl_u8_x(__VA_ARGS__) -#define svqshl_u32_x(...) __builtin_sve_svqshl_u32_x(__VA_ARGS__) -#define svqshl_u64_x(...) __builtin_sve_svqshl_u64_x(__VA_ARGS__) -#define svqshl_u16_x(...) __builtin_sve_svqshl_u16_x(__VA_ARGS__) -#define svqshl_u8_z(...) __builtin_sve_svqshl_u8_z(__VA_ARGS__) -#define svqshl_u32_z(...) __builtin_sve_svqshl_u32_z(__VA_ARGS__) -#define svqshl_u64_z(...) __builtin_sve_svqshl_u64_z(__VA_ARGS__) -#define svqshl_u16_z(...) __builtin_sve_svqshl_u16_z(__VA_ARGS__) -#define svqshlu_n_s8_m(...) __builtin_sve_svqshlu_n_s8_m(__VA_ARGS__) -#define svqshlu_n_s32_m(...) __builtin_sve_svqshlu_n_s32_m(__VA_ARGS__) -#define svqshlu_n_s64_m(...) __builtin_sve_svqshlu_n_s64_m(__VA_ARGS__) -#define svqshlu_n_s16_m(...) __builtin_sve_svqshlu_n_s16_m(__VA_ARGS__) -#define svqshlu_n_s8_x(...) __builtin_sve_svqshlu_n_s8_x(__VA_ARGS__) -#define svqshlu_n_s32_x(...) __builtin_sve_svqshlu_n_s32_x(__VA_ARGS__) -#define svqshlu_n_s64_x(...) __builtin_sve_svqshlu_n_s64_x(__VA_ARGS__) -#define svqshlu_n_s16_x(...) __builtin_sve_svqshlu_n_s16_x(__VA_ARGS__) -#define svqshlu_n_s8_z(...) __builtin_sve_svqshlu_n_s8_z(__VA_ARGS__) -#define svqshlu_n_s32_z(...) __builtin_sve_svqshlu_n_s32_z(__VA_ARGS__) -#define svqshlu_n_s64_z(...) __builtin_sve_svqshlu_n_s64_z(__VA_ARGS__) -#define svqshlu_n_s16_z(...) __builtin_sve_svqshlu_n_s16_z(__VA_ARGS__) -#define svqshrnb_n_s32(...) __builtin_sve_svqshrnb_n_s32(__VA_ARGS__) -#define svqshrnb_n_s64(...) __builtin_sve_svqshrnb_n_s64(__VA_ARGS__) -#define svqshrnb_n_s16(...) __builtin_sve_svqshrnb_n_s16(__VA_ARGS__) -#define svqshrnb_n_u32(...) __builtin_sve_svqshrnb_n_u32(__VA_ARGS__) -#define svqshrnb_n_u64(...) __builtin_sve_svqshrnb_n_u64(__VA_ARGS__) -#define svqshrnb_n_u16(...) __builtin_sve_svqshrnb_n_u16(__VA_ARGS__) -#define svqshrnt_n_s32(...) __builtin_sve_svqshrnt_n_s32(__VA_ARGS__) -#define svqshrnt_n_s64(...) __builtin_sve_svqshrnt_n_s64(__VA_ARGS__) -#define svqshrnt_n_s16(...) __builtin_sve_svqshrnt_n_s16(__VA_ARGS__) -#define svqshrnt_n_u32(...) __builtin_sve_svqshrnt_n_u32(__VA_ARGS__) -#define svqshrnt_n_u64(...) __builtin_sve_svqshrnt_n_u64(__VA_ARGS__) -#define svqshrnt_n_u16(...) __builtin_sve_svqshrnt_n_u16(__VA_ARGS__) -#define svqshrunb_n_s32(...) __builtin_sve_svqshrunb_n_s32(__VA_ARGS__) -#define svqshrunb_n_s64(...) __builtin_sve_svqshrunb_n_s64(__VA_ARGS__) -#define svqshrunb_n_s16(...) __builtin_sve_svqshrunb_n_s16(__VA_ARGS__) -#define svqshrunt_n_s32(...) __builtin_sve_svqshrunt_n_s32(__VA_ARGS__) -#define svqshrunt_n_s64(...) __builtin_sve_svqshrunt_n_s64(__VA_ARGS__) -#define svqshrunt_n_s16(...) __builtin_sve_svqshrunt_n_s16(__VA_ARGS__) -#define svqsub_n_s8_m(...) __builtin_sve_svqsub_n_s8_m(__VA_ARGS__) -#define svqsub_n_s32_m(...) __builtin_sve_svqsub_n_s32_m(__VA_ARGS__) -#define svqsub_n_s64_m(...) __builtin_sve_svqsub_n_s64_m(__VA_ARGS__) -#define svqsub_n_s16_m(...) __builtin_sve_svqsub_n_s16_m(__VA_ARGS__) -#define svqsub_n_s8_x(...) __builtin_sve_svqsub_n_s8_x(__VA_ARGS__) -#define svqsub_n_s32_x(...) __builtin_sve_svqsub_n_s32_x(__VA_ARGS__) -#define svqsub_n_s64_x(...) __builtin_sve_svqsub_n_s64_x(__VA_ARGS__) -#define svqsub_n_s16_x(...) __builtin_sve_svqsub_n_s16_x(__VA_ARGS__) -#define svqsub_n_s8_z(...) __builtin_sve_svqsub_n_s8_z(__VA_ARGS__) -#define svqsub_n_s32_z(...) __builtin_sve_svqsub_n_s32_z(__VA_ARGS__) -#define svqsub_n_s64_z(...) __builtin_sve_svqsub_n_s64_z(__VA_ARGS__) -#define svqsub_n_s16_z(...) __builtin_sve_svqsub_n_s16_z(__VA_ARGS__) -#define svqsub_n_u8_m(...) __builtin_sve_svqsub_n_u8_m(__VA_ARGS__) -#define svqsub_n_u32_m(...) __builtin_sve_svqsub_n_u32_m(__VA_ARGS__) -#define svqsub_n_u64_m(...) __builtin_sve_svqsub_n_u64_m(__VA_ARGS__) -#define svqsub_n_u16_m(...) __builtin_sve_svqsub_n_u16_m(__VA_ARGS__) -#define svqsub_n_u8_x(...) __builtin_sve_svqsub_n_u8_x(__VA_ARGS__) -#define svqsub_n_u32_x(...) __builtin_sve_svqsub_n_u32_x(__VA_ARGS__) -#define svqsub_n_u64_x(...) __builtin_sve_svqsub_n_u64_x(__VA_ARGS__) -#define svqsub_n_u16_x(...) __builtin_sve_svqsub_n_u16_x(__VA_ARGS__) -#define svqsub_n_u8_z(...) __builtin_sve_svqsub_n_u8_z(__VA_ARGS__) -#define svqsub_n_u32_z(...) __builtin_sve_svqsub_n_u32_z(__VA_ARGS__) -#define svqsub_n_u64_z(...) __builtin_sve_svqsub_n_u64_z(__VA_ARGS__) -#define svqsub_n_u16_z(...) __builtin_sve_svqsub_n_u16_z(__VA_ARGS__) -#define svqsub_s8_m(...) __builtin_sve_svqsub_s8_m(__VA_ARGS__) -#define svqsub_s32_m(...) __builtin_sve_svqsub_s32_m(__VA_ARGS__) -#define svqsub_s64_m(...) __builtin_sve_svqsub_s64_m(__VA_ARGS__) -#define svqsub_s16_m(...) __builtin_sve_svqsub_s16_m(__VA_ARGS__) -#define svqsub_s8_x(...) __builtin_sve_svqsub_s8_x(__VA_ARGS__) -#define svqsub_s32_x(...) __builtin_sve_svqsub_s32_x(__VA_ARGS__) -#define svqsub_s64_x(...) __builtin_sve_svqsub_s64_x(__VA_ARGS__) -#define svqsub_s16_x(...) __builtin_sve_svqsub_s16_x(__VA_ARGS__) -#define svqsub_s8_z(...) __builtin_sve_svqsub_s8_z(__VA_ARGS__) -#define svqsub_s32_z(...) __builtin_sve_svqsub_s32_z(__VA_ARGS__) -#define svqsub_s64_z(...) __builtin_sve_svqsub_s64_z(__VA_ARGS__) -#define svqsub_s16_z(...) __builtin_sve_svqsub_s16_z(__VA_ARGS__) -#define svqsub_u8_m(...) __builtin_sve_svqsub_u8_m(__VA_ARGS__) -#define svqsub_u32_m(...) __builtin_sve_svqsub_u32_m(__VA_ARGS__) -#define svqsub_u64_m(...) __builtin_sve_svqsub_u64_m(__VA_ARGS__) -#define svqsub_u16_m(...) __builtin_sve_svqsub_u16_m(__VA_ARGS__) -#define svqsub_u8_x(...) __builtin_sve_svqsub_u8_x(__VA_ARGS__) -#define svqsub_u32_x(...) __builtin_sve_svqsub_u32_x(__VA_ARGS__) -#define svqsub_u64_x(...) __builtin_sve_svqsub_u64_x(__VA_ARGS__) -#define svqsub_u16_x(...) __builtin_sve_svqsub_u16_x(__VA_ARGS__) -#define svqsub_u8_z(...) __builtin_sve_svqsub_u8_z(__VA_ARGS__) -#define svqsub_u32_z(...) __builtin_sve_svqsub_u32_z(__VA_ARGS__) -#define svqsub_u64_z(...) __builtin_sve_svqsub_u64_z(__VA_ARGS__) -#define svqsub_u16_z(...) __builtin_sve_svqsub_u16_z(__VA_ARGS__) -#define svqsubr_n_s8_m(...) __builtin_sve_svqsubr_n_s8_m(__VA_ARGS__) -#define svqsubr_n_s32_m(...) __builtin_sve_svqsubr_n_s32_m(__VA_ARGS__) -#define svqsubr_n_s64_m(...) __builtin_sve_svqsubr_n_s64_m(__VA_ARGS__) -#define svqsubr_n_s16_m(...) __builtin_sve_svqsubr_n_s16_m(__VA_ARGS__) -#define svqsubr_n_s8_x(...) __builtin_sve_svqsubr_n_s8_x(__VA_ARGS__) -#define svqsubr_n_s32_x(...) __builtin_sve_svqsubr_n_s32_x(__VA_ARGS__) -#define svqsubr_n_s64_x(...) __builtin_sve_svqsubr_n_s64_x(__VA_ARGS__) -#define svqsubr_n_s16_x(...) __builtin_sve_svqsubr_n_s16_x(__VA_ARGS__) -#define svqsubr_n_s8_z(...) __builtin_sve_svqsubr_n_s8_z(__VA_ARGS__) -#define svqsubr_n_s32_z(...) __builtin_sve_svqsubr_n_s32_z(__VA_ARGS__) -#define svqsubr_n_s64_z(...) __builtin_sve_svqsubr_n_s64_z(__VA_ARGS__) -#define svqsubr_n_s16_z(...) __builtin_sve_svqsubr_n_s16_z(__VA_ARGS__) -#define svqsubr_n_u8_m(...) __builtin_sve_svqsubr_n_u8_m(__VA_ARGS__) -#define svqsubr_n_u32_m(...) __builtin_sve_svqsubr_n_u32_m(__VA_ARGS__) -#define svqsubr_n_u64_m(...) __builtin_sve_svqsubr_n_u64_m(__VA_ARGS__) -#define svqsubr_n_u16_m(...) __builtin_sve_svqsubr_n_u16_m(__VA_ARGS__) -#define svqsubr_n_u8_x(...) __builtin_sve_svqsubr_n_u8_x(__VA_ARGS__) -#define svqsubr_n_u32_x(...) __builtin_sve_svqsubr_n_u32_x(__VA_ARGS__) -#define svqsubr_n_u64_x(...) __builtin_sve_svqsubr_n_u64_x(__VA_ARGS__) -#define svqsubr_n_u16_x(...) __builtin_sve_svqsubr_n_u16_x(__VA_ARGS__) -#define svqsubr_n_u8_z(...) __builtin_sve_svqsubr_n_u8_z(__VA_ARGS__) -#define svqsubr_n_u32_z(...) __builtin_sve_svqsubr_n_u32_z(__VA_ARGS__) -#define svqsubr_n_u64_z(...) __builtin_sve_svqsubr_n_u64_z(__VA_ARGS__) -#define svqsubr_n_u16_z(...) __builtin_sve_svqsubr_n_u16_z(__VA_ARGS__) -#define svqsubr_s8_m(...) __builtin_sve_svqsubr_s8_m(__VA_ARGS__) -#define svqsubr_s32_m(...) __builtin_sve_svqsubr_s32_m(__VA_ARGS__) -#define svqsubr_s64_m(...) __builtin_sve_svqsubr_s64_m(__VA_ARGS__) -#define svqsubr_s16_m(...) __builtin_sve_svqsubr_s16_m(__VA_ARGS__) -#define svqsubr_s8_x(...) __builtin_sve_svqsubr_s8_x(__VA_ARGS__) -#define svqsubr_s32_x(...) __builtin_sve_svqsubr_s32_x(__VA_ARGS__) -#define svqsubr_s64_x(...) __builtin_sve_svqsubr_s64_x(__VA_ARGS__) -#define svqsubr_s16_x(...) __builtin_sve_svqsubr_s16_x(__VA_ARGS__) -#define svqsubr_s8_z(...) __builtin_sve_svqsubr_s8_z(__VA_ARGS__) -#define svqsubr_s32_z(...) __builtin_sve_svqsubr_s32_z(__VA_ARGS__) -#define svqsubr_s64_z(...) __builtin_sve_svqsubr_s64_z(__VA_ARGS__) -#define svqsubr_s16_z(...) __builtin_sve_svqsubr_s16_z(__VA_ARGS__) -#define svqsubr_u8_m(...) __builtin_sve_svqsubr_u8_m(__VA_ARGS__) -#define svqsubr_u32_m(...) __builtin_sve_svqsubr_u32_m(__VA_ARGS__) -#define svqsubr_u64_m(...) __builtin_sve_svqsubr_u64_m(__VA_ARGS__) -#define svqsubr_u16_m(...) __builtin_sve_svqsubr_u16_m(__VA_ARGS__) -#define svqsubr_u8_x(...) __builtin_sve_svqsubr_u8_x(__VA_ARGS__) -#define svqsubr_u32_x(...) __builtin_sve_svqsubr_u32_x(__VA_ARGS__) -#define svqsubr_u64_x(...) __builtin_sve_svqsubr_u64_x(__VA_ARGS__) -#define svqsubr_u16_x(...) __builtin_sve_svqsubr_u16_x(__VA_ARGS__) -#define svqsubr_u8_z(...) __builtin_sve_svqsubr_u8_z(__VA_ARGS__) -#define svqsubr_u32_z(...) __builtin_sve_svqsubr_u32_z(__VA_ARGS__) -#define svqsubr_u64_z(...) __builtin_sve_svqsubr_u64_z(__VA_ARGS__) -#define svqsubr_u16_z(...) __builtin_sve_svqsubr_u16_z(__VA_ARGS__) -#define svqxtnb_s32(...) __builtin_sve_svqxtnb_s32(__VA_ARGS__) -#define svqxtnb_s64(...) __builtin_sve_svqxtnb_s64(__VA_ARGS__) -#define svqxtnb_s16(...) __builtin_sve_svqxtnb_s16(__VA_ARGS__) -#define svqxtnb_u32(...) __builtin_sve_svqxtnb_u32(__VA_ARGS__) -#define svqxtnb_u64(...) __builtin_sve_svqxtnb_u64(__VA_ARGS__) -#define svqxtnb_u16(...) __builtin_sve_svqxtnb_u16(__VA_ARGS__) -#define svqxtnt_s32(...) __builtin_sve_svqxtnt_s32(__VA_ARGS__) -#define svqxtnt_s64(...) __builtin_sve_svqxtnt_s64(__VA_ARGS__) -#define svqxtnt_s16(...) __builtin_sve_svqxtnt_s16(__VA_ARGS__) -#define svqxtnt_u32(...) __builtin_sve_svqxtnt_u32(__VA_ARGS__) -#define svqxtnt_u64(...) __builtin_sve_svqxtnt_u64(__VA_ARGS__) -#define svqxtnt_u16(...) __builtin_sve_svqxtnt_u16(__VA_ARGS__) -#define svqxtunb_s32(...) __builtin_sve_svqxtunb_s32(__VA_ARGS__) -#define svqxtunb_s64(...) __builtin_sve_svqxtunb_s64(__VA_ARGS__) -#define svqxtunb_s16(...) __builtin_sve_svqxtunb_s16(__VA_ARGS__) -#define svqxtunt_s32(...) __builtin_sve_svqxtunt_s32(__VA_ARGS__) -#define svqxtunt_s64(...) __builtin_sve_svqxtunt_s64(__VA_ARGS__) -#define svqxtunt_s16(...) __builtin_sve_svqxtunt_s16(__VA_ARGS__) -#define svraddhnb_n_u32(...) __builtin_sve_svraddhnb_n_u32(__VA_ARGS__) -#define svraddhnb_n_u64(...) __builtin_sve_svraddhnb_n_u64(__VA_ARGS__) -#define svraddhnb_n_u16(...) __builtin_sve_svraddhnb_n_u16(__VA_ARGS__) -#define svraddhnb_n_s32(...) __builtin_sve_svraddhnb_n_s32(__VA_ARGS__) -#define svraddhnb_n_s64(...) __builtin_sve_svraddhnb_n_s64(__VA_ARGS__) -#define svraddhnb_n_s16(...) __builtin_sve_svraddhnb_n_s16(__VA_ARGS__) -#define svraddhnb_u32(...) __builtin_sve_svraddhnb_u32(__VA_ARGS__) -#define svraddhnb_u64(...) __builtin_sve_svraddhnb_u64(__VA_ARGS__) -#define svraddhnb_u16(...) __builtin_sve_svraddhnb_u16(__VA_ARGS__) -#define svraddhnb_s32(...) __builtin_sve_svraddhnb_s32(__VA_ARGS__) -#define svraddhnb_s64(...) __builtin_sve_svraddhnb_s64(__VA_ARGS__) -#define svraddhnb_s16(...) __builtin_sve_svraddhnb_s16(__VA_ARGS__) -#define svraddhnt_n_u32(...) __builtin_sve_svraddhnt_n_u32(__VA_ARGS__) -#define svraddhnt_n_u64(...) __builtin_sve_svraddhnt_n_u64(__VA_ARGS__) -#define svraddhnt_n_u16(...) __builtin_sve_svraddhnt_n_u16(__VA_ARGS__) -#define svraddhnt_n_s32(...) __builtin_sve_svraddhnt_n_s32(__VA_ARGS__) -#define svraddhnt_n_s64(...) __builtin_sve_svraddhnt_n_s64(__VA_ARGS__) -#define svraddhnt_n_s16(...) __builtin_sve_svraddhnt_n_s16(__VA_ARGS__) -#define svraddhnt_u32(...) __builtin_sve_svraddhnt_u32(__VA_ARGS__) -#define svraddhnt_u64(...) __builtin_sve_svraddhnt_u64(__VA_ARGS__) -#define svraddhnt_u16(...) __builtin_sve_svraddhnt_u16(__VA_ARGS__) -#define svraddhnt_s32(...) __builtin_sve_svraddhnt_s32(__VA_ARGS__) -#define svraddhnt_s64(...) __builtin_sve_svraddhnt_s64(__VA_ARGS__) -#define svraddhnt_s16(...) __builtin_sve_svraddhnt_s16(__VA_ARGS__) -#define svrecpe_u32_m(...) __builtin_sve_svrecpe_u32_m(__VA_ARGS__) -#define svrecpe_u32_x(...) __builtin_sve_svrecpe_u32_x(__VA_ARGS__) -#define svrecpe_u32_z(...) __builtin_sve_svrecpe_u32_z(__VA_ARGS__) -#define svrhadd_n_s8_m(...) __builtin_sve_svrhadd_n_s8_m(__VA_ARGS__) -#define svrhadd_n_s32_m(...) __builtin_sve_svrhadd_n_s32_m(__VA_ARGS__) -#define svrhadd_n_s64_m(...) __builtin_sve_svrhadd_n_s64_m(__VA_ARGS__) -#define svrhadd_n_s16_m(...) __builtin_sve_svrhadd_n_s16_m(__VA_ARGS__) -#define svrhadd_n_s8_x(...) __builtin_sve_svrhadd_n_s8_x(__VA_ARGS__) -#define svrhadd_n_s32_x(...) __builtin_sve_svrhadd_n_s32_x(__VA_ARGS__) -#define svrhadd_n_s64_x(...) __builtin_sve_svrhadd_n_s64_x(__VA_ARGS__) -#define svrhadd_n_s16_x(...) __builtin_sve_svrhadd_n_s16_x(__VA_ARGS__) -#define svrhadd_n_s8_z(...) __builtin_sve_svrhadd_n_s8_z(__VA_ARGS__) -#define svrhadd_n_s32_z(...) __builtin_sve_svrhadd_n_s32_z(__VA_ARGS__) -#define svrhadd_n_s64_z(...) __builtin_sve_svrhadd_n_s64_z(__VA_ARGS__) -#define svrhadd_n_s16_z(...) __builtin_sve_svrhadd_n_s16_z(__VA_ARGS__) -#define svrhadd_n_u8_m(...) __builtin_sve_svrhadd_n_u8_m(__VA_ARGS__) -#define svrhadd_n_u32_m(...) __builtin_sve_svrhadd_n_u32_m(__VA_ARGS__) -#define svrhadd_n_u64_m(...) __builtin_sve_svrhadd_n_u64_m(__VA_ARGS__) -#define svrhadd_n_u16_m(...) __builtin_sve_svrhadd_n_u16_m(__VA_ARGS__) -#define svrhadd_n_u8_x(...) __builtin_sve_svrhadd_n_u8_x(__VA_ARGS__) -#define svrhadd_n_u32_x(...) __builtin_sve_svrhadd_n_u32_x(__VA_ARGS__) -#define svrhadd_n_u64_x(...) __builtin_sve_svrhadd_n_u64_x(__VA_ARGS__) -#define svrhadd_n_u16_x(...) __builtin_sve_svrhadd_n_u16_x(__VA_ARGS__) -#define svrhadd_n_u8_z(...) __builtin_sve_svrhadd_n_u8_z(__VA_ARGS__) -#define svrhadd_n_u32_z(...) __builtin_sve_svrhadd_n_u32_z(__VA_ARGS__) -#define svrhadd_n_u64_z(...) __builtin_sve_svrhadd_n_u64_z(__VA_ARGS__) -#define svrhadd_n_u16_z(...) __builtin_sve_svrhadd_n_u16_z(__VA_ARGS__) -#define svrhadd_s8_m(...) __builtin_sve_svrhadd_s8_m(__VA_ARGS__) -#define svrhadd_s32_m(...) __builtin_sve_svrhadd_s32_m(__VA_ARGS__) -#define svrhadd_s64_m(...) __builtin_sve_svrhadd_s64_m(__VA_ARGS__) -#define svrhadd_s16_m(...) __builtin_sve_svrhadd_s16_m(__VA_ARGS__) -#define svrhadd_s8_x(...) __builtin_sve_svrhadd_s8_x(__VA_ARGS__) -#define svrhadd_s32_x(...) __builtin_sve_svrhadd_s32_x(__VA_ARGS__) -#define svrhadd_s64_x(...) __builtin_sve_svrhadd_s64_x(__VA_ARGS__) -#define svrhadd_s16_x(...) __builtin_sve_svrhadd_s16_x(__VA_ARGS__) -#define svrhadd_s8_z(...) __builtin_sve_svrhadd_s8_z(__VA_ARGS__) -#define svrhadd_s32_z(...) __builtin_sve_svrhadd_s32_z(__VA_ARGS__) -#define svrhadd_s64_z(...) __builtin_sve_svrhadd_s64_z(__VA_ARGS__) -#define svrhadd_s16_z(...) __builtin_sve_svrhadd_s16_z(__VA_ARGS__) -#define svrhadd_u8_m(...) __builtin_sve_svrhadd_u8_m(__VA_ARGS__) -#define svrhadd_u32_m(...) __builtin_sve_svrhadd_u32_m(__VA_ARGS__) -#define svrhadd_u64_m(...) __builtin_sve_svrhadd_u64_m(__VA_ARGS__) -#define svrhadd_u16_m(...) __builtin_sve_svrhadd_u16_m(__VA_ARGS__) -#define svrhadd_u8_x(...) __builtin_sve_svrhadd_u8_x(__VA_ARGS__) -#define svrhadd_u32_x(...) __builtin_sve_svrhadd_u32_x(__VA_ARGS__) -#define svrhadd_u64_x(...) __builtin_sve_svrhadd_u64_x(__VA_ARGS__) -#define svrhadd_u16_x(...) __builtin_sve_svrhadd_u16_x(__VA_ARGS__) -#define svrhadd_u8_z(...) __builtin_sve_svrhadd_u8_z(__VA_ARGS__) -#define svrhadd_u32_z(...) __builtin_sve_svrhadd_u32_z(__VA_ARGS__) -#define svrhadd_u64_z(...) __builtin_sve_svrhadd_u64_z(__VA_ARGS__) -#define svrhadd_u16_z(...) __builtin_sve_svrhadd_u16_z(__VA_ARGS__) -#define svrshl_n_s8_m(...) __builtin_sve_svrshl_n_s8_m(__VA_ARGS__) -#define svrshl_n_s32_m(...) __builtin_sve_svrshl_n_s32_m(__VA_ARGS__) -#define svrshl_n_s64_m(...) __builtin_sve_svrshl_n_s64_m(__VA_ARGS__) -#define svrshl_n_s16_m(...) __builtin_sve_svrshl_n_s16_m(__VA_ARGS__) -#define svrshl_n_s8_x(...) __builtin_sve_svrshl_n_s8_x(__VA_ARGS__) -#define svrshl_n_s32_x(...) __builtin_sve_svrshl_n_s32_x(__VA_ARGS__) -#define svrshl_n_s64_x(...) __builtin_sve_svrshl_n_s64_x(__VA_ARGS__) -#define svrshl_n_s16_x(...) __builtin_sve_svrshl_n_s16_x(__VA_ARGS__) -#define svrshl_n_s8_z(...) __builtin_sve_svrshl_n_s8_z(__VA_ARGS__) -#define svrshl_n_s32_z(...) __builtin_sve_svrshl_n_s32_z(__VA_ARGS__) -#define svrshl_n_s64_z(...) __builtin_sve_svrshl_n_s64_z(__VA_ARGS__) -#define svrshl_n_s16_z(...) __builtin_sve_svrshl_n_s16_z(__VA_ARGS__) -#define svrshl_n_u8_m(...) __builtin_sve_svrshl_n_u8_m(__VA_ARGS__) -#define svrshl_n_u32_m(...) __builtin_sve_svrshl_n_u32_m(__VA_ARGS__) -#define svrshl_n_u64_m(...) __builtin_sve_svrshl_n_u64_m(__VA_ARGS__) -#define svrshl_n_u16_m(...) __builtin_sve_svrshl_n_u16_m(__VA_ARGS__) -#define svrshl_n_u8_x(...) __builtin_sve_svrshl_n_u8_x(__VA_ARGS__) -#define svrshl_n_u32_x(...) __builtin_sve_svrshl_n_u32_x(__VA_ARGS__) -#define svrshl_n_u64_x(...) __builtin_sve_svrshl_n_u64_x(__VA_ARGS__) -#define svrshl_n_u16_x(...) __builtin_sve_svrshl_n_u16_x(__VA_ARGS__) -#define svrshl_n_u8_z(...) __builtin_sve_svrshl_n_u8_z(__VA_ARGS__) -#define svrshl_n_u32_z(...) __builtin_sve_svrshl_n_u32_z(__VA_ARGS__) -#define svrshl_n_u64_z(...) __builtin_sve_svrshl_n_u64_z(__VA_ARGS__) -#define svrshl_n_u16_z(...) __builtin_sve_svrshl_n_u16_z(__VA_ARGS__) -#define svrshl_s8_m(...) __builtin_sve_svrshl_s8_m(__VA_ARGS__) -#define svrshl_s32_m(...) __builtin_sve_svrshl_s32_m(__VA_ARGS__) -#define svrshl_s64_m(...) __builtin_sve_svrshl_s64_m(__VA_ARGS__) -#define svrshl_s16_m(...) __builtin_sve_svrshl_s16_m(__VA_ARGS__) -#define svrshl_s8_x(...) __builtin_sve_svrshl_s8_x(__VA_ARGS__) -#define svrshl_s32_x(...) __builtin_sve_svrshl_s32_x(__VA_ARGS__) -#define svrshl_s64_x(...) __builtin_sve_svrshl_s64_x(__VA_ARGS__) -#define svrshl_s16_x(...) __builtin_sve_svrshl_s16_x(__VA_ARGS__) -#define svrshl_s8_z(...) __builtin_sve_svrshl_s8_z(__VA_ARGS__) -#define svrshl_s32_z(...) __builtin_sve_svrshl_s32_z(__VA_ARGS__) -#define svrshl_s64_z(...) __builtin_sve_svrshl_s64_z(__VA_ARGS__) -#define svrshl_s16_z(...) __builtin_sve_svrshl_s16_z(__VA_ARGS__) -#define svrshl_u8_m(...) __builtin_sve_svrshl_u8_m(__VA_ARGS__) -#define svrshl_u32_m(...) __builtin_sve_svrshl_u32_m(__VA_ARGS__) -#define svrshl_u64_m(...) __builtin_sve_svrshl_u64_m(__VA_ARGS__) -#define svrshl_u16_m(...) __builtin_sve_svrshl_u16_m(__VA_ARGS__) -#define svrshl_u8_x(...) __builtin_sve_svrshl_u8_x(__VA_ARGS__) -#define svrshl_u32_x(...) __builtin_sve_svrshl_u32_x(__VA_ARGS__) -#define svrshl_u64_x(...) __builtin_sve_svrshl_u64_x(__VA_ARGS__) -#define svrshl_u16_x(...) __builtin_sve_svrshl_u16_x(__VA_ARGS__) -#define svrshl_u8_z(...) __builtin_sve_svrshl_u8_z(__VA_ARGS__) -#define svrshl_u32_z(...) __builtin_sve_svrshl_u32_z(__VA_ARGS__) -#define svrshl_u64_z(...) __builtin_sve_svrshl_u64_z(__VA_ARGS__) -#define svrshl_u16_z(...) __builtin_sve_svrshl_u16_z(__VA_ARGS__) -#define svrshr_n_s8_m(...) __builtin_sve_svrshr_n_s8_m(__VA_ARGS__) -#define svrshr_n_s32_m(...) __builtin_sve_svrshr_n_s32_m(__VA_ARGS__) -#define svrshr_n_s64_m(...) __builtin_sve_svrshr_n_s64_m(__VA_ARGS__) -#define svrshr_n_s16_m(...) __builtin_sve_svrshr_n_s16_m(__VA_ARGS__) -#define svrshr_n_u8_m(...) __builtin_sve_svrshr_n_u8_m(__VA_ARGS__) -#define svrshr_n_u32_m(...) __builtin_sve_svrshr_n_u32_m(__VA_ARGS__) -#define svrshr_n_u64_m(...) __builtin_sve_svrshr_n_u64_m(__VA_ARGS__) -#define svrshr_n_u16_m(...) __builtin_sve_svrshr_n_u16_m(__VA_ARGS__) -#define svrshr_n_s8_x(...) __builtin_sve_svrshr_n_s8_x(__VA_ARGS__) -#define svrshr_n_s32_x(...) __builtin_sve_svrshr_n_s32_x(__VA_ARGS__) -#define svrshr_n_s64_x(...) __builtin_sve_svrshr_n_s64_x(__VA_ARGS__) -#define svrshr_n_s16_x(...) __builtin_sve_svrshr_n_s16_x(__VA_ARGS__) -#define svrshr_n_u8_x(...) __builtin_sve_svrshr_n_u8_x(__VA_ARGS__) -#define svrshr_n_u32_x(...) __builtin_sve_svrshr_n_u32_x(__VA_ARGS__) -#define svrshr_n_u64_x(...) __builtin_sve_svrshr_n_u64_x(__VA_ARGS__) -#define svrshr_n_u16_x(...) __builtin_sve_svrshr_n_u16_x(__VA_ARGS__) -#define svrshr_n_s8_z(...) __builtin_sve_svrshr_n_s8_z(__VA_ARGS__) -#define svrshr_n_s32_z(...) __builtin_sve_svrshr_n_s32_z(__VA_ARGS__) -#define svrshr_n_s64_z(...) __builtin_sve_svrshr_n_s64_z(__VA_ARGS__) -#define svrshr_n_s16_z(...) __builtin_sve_svrshr_n_s16_z(__VA_ARGS__) -#define svrshr_n_u8_z(...) __builtin_sve_svrshr_n_u8_z(__VA_ARGS__) -#define svrshr_n_u32_z(...) __builtin_sve_svrshr_n_u32_z(__VA_ARGS__) -#define svrshr_n_u64_z(...) __builtin_sve_svrshr_n_u64_z(__VA_ARGS__) -#define svrshr_n_u16_z(...) __builtin_sve_svrshr_n_u16_z(__VA_ARGS__) -#define svrshrnb_n_u32(...) __builtin_sve_svrshrnb_n_u32(__VA_ARGS__) -#define svrshrnb_n_u64(...) __builtin_sve_svrshrnb_n_u64(__VA_ARGS__) -#define svrshrnb_n_u16(...) __builtin_sve_svrshrnb_n_u16(__VA_ARGS__) -#define svrshrnb_n_s32(...) __builtin_sve_svrshrnb_n_s32(__VA_ARGS__) -#define svrshrnb_n_s64(...) __builtin_sve_svrshrnb_n_s64(__VA_ARGS__) -#define svrshrnb_n_s16(...) __builtin_sve_svrshrnb_n_s16(__VA_ARGS__) -#define svrshrnt_n_u32(...) __builtin_sve_svrshrnt_n_u32(__VA_ARGS__) -#define svrshrnt_n_u64(...) __builtin_sve_svrshrnt_n_u64(__VA_ARGS__) -#define svrshrnt_n_u16(...) __builtin_sve_svrshrnt_n_u16(__VA_ARGS__) -#define svrshrnt_n_s32(...) __builtin_sve_svrshrnt_n_s32(__VA_ARGS__) -#define svrshrnt_n_s64(...) __builtin_sve_svrshrnt_n_s64(__VA_ARGS__) -#define svrshrnt_n_s16(...) __builtin_sve_svrshrnt_n_s16(__VA_ARGS__) -#define svrsqrte_u32_m(...) __builtin_sve_svrsqrte_u32_m(__VA_ARGS__) -#define svrsqrte_u32_x(...) __builtin_sve_svrsqrte_u32_x(__VA_ARGS__) -#define svrsqrte_u32_z(...) __builtin_sve_svrsqrte_u32_z(__VA_ARGS__) -#define svrsra_n_s8(...) __builtin_sve_svrsra_n_s8(__VA_ARGS__) -#define svrsra_n_s32(...) __builtin_sve_svrsra_n_s32(__VA_ARGS__) -#define svrsra_n_s64(...) __builtin_sve_svrsra_n_s64(__VA_ARGS__) -#define svrsra_n_s16(...) __builtin_sve_svrsra_n_s16(__VA_ARGS__) -#define svrsra_n_u8(...) __builtin_sve_svrsra_n_u8(__VA_ARGS__) -#define svrsra_n_u32(...) __builtin_sve_svrsra_n_u32(__VA_ARGS__) -#define svrsra_n_u64(...) __builtin_sve_svrsra_n_u64(__VA_ARGS__) -#define svrsra_n_u16(...) __builtin_sve_svrsra_n_u16(__VA_ARGS__) -#define svrsubhnb_n_u32(...) __builtin_sve_svrsubhnb_n_u32(__VA_ARGS__) -#define svrsubhnb_n_u64(...) __builtin_sve_svrsubhnb_n_u64(__VA_ARGS__) -#define svrsubhnb_n_u16(...) __builtin_sve_svrsubhnb_n_u16(__VA_ARGS__) -#define svrsubhnb_n_s32(...) __builtin_sve_svrsubhnb_n_s32(__VA_ARGS__) -#define svrsubhnb_n_s64(...) __builtin_sve_svrsubhnb_n_s64(__VA_ARGS__) -#define svrsubhnb_n_s16(...) __builtin_sve_svrsubhnb_n_s16(__VA_ARGS__) -#define svrsubhnb_u32(...) __builtin_sve_svrsubhnb_u32(__VA_ARGS__) -#define svrsubhnb_u64(...) __builtin_sve_svrsubhnb_u64(__VA_ARGS__) -#define svrsubhnb_u16(...) __builtin_sve_svrsubhnb_u16(__VA_ARGS__) -#define svrsubhnb_s32(...) __builtin_sve_svrsubhnb_s32(__VA_ARGS__) -#define svrsubhnb_s64(...) __builtin_sve_svrsubhnb_s64(__VA_ARGS__) -#define svrsubhnb_s16(...) __builtin_sve_svrsubhnb_s16(__VA_ARGS__) -#define svrsubhnt_n_u32(...) __builtin_sve_svrsubhnt_n_u32(__VA_ARGS__) -#define svrsubhnt_n_u64(...) __builtin_sve_svrsubhnt_n_u64(__VA_ARGS__) -#define svrsubhnt_n_u16(...) __builtin_sve_svrsubhnt_n_u16(__VA_ARGS__) -#define svrsubhnt_n_s32(...) __builtin_sve_svrsubhnt_n_s32(__VA_ARGS__) -#define svrsubhnt_n_s64(...) __builtin_sve_svrsubhnt_n_s64(__VA_ARGS__) -#define svrsubhnt_n_s16(...) __builtin_sve_svrsubhnt_n_s16(__VA_ARGS__) -#define svrsubhnt_u32(...) __builtin_sve_svrsubhnt_u32(__VA_ARGS__) -#define svrsubhnt_u64(...) __builtin_sve_svrsubhnt_u64(__VA_ARGS__) -#define svrsubhnt_u16(...) __builtin_sve_svrsubhnt_u16(__VA_ARGS__) -#define svrsubhnt_s32(...) __builtin_sve_svrsubhnt_s32(__VA_ARGS__) -#define svrsubhnt_s64(...) __builtin_sve_svrsubhnt_s64(__VA_ARGS__) -#define svrsubhnt_s16(...) __builtin_sve_svrsubhnt_s16(__VA_ARGS__) -#define svsbclb_n_u32(...) __builtin_sve_svsbclb_n_u32(__VA_ARGS__) -#define svsbclb_n_u64(...) __builtin_sve_svsbclb_n_u64(__VA_ARGS__) -#define svsbclb_u32(...) __builtin_sve_svsbclb_u32(__VA_ARGS__) -#define svsbclb_u64(...) __builtin_sve_svsbclb_u64(__VA_ARGS__) -#define svsbclt_n_u32(...) __builtin_sve_svsbclt_n_u32(__VA_ARGS__) -#define svsbclt_n_u64(...) __builtin_sve_svsbclt_n_u64(__VA_ARGS__) -#define svsbclt_u32(...) __builtin_sve_svsbclt_u32(__VA_ARGS__) -#define svsbclt_u64(...) __builtin_sve_svsbclt_u64(__VA_ARGS__) -#define svshllb_n_s32(...) __builtin_sve_svshllb_n_s32(__VA_ARGS__) -#define svshllb_n_s64(...) __builtin_sve_svshllb_n_s64(__VA_ARGS__) -#define svshllb_n_s16(...) __builtin_sve_svshllb_n_s16(__VA_ARGS__) -#define svshllb_n_u32(...) __builtin_sve_svshllb_n_u32(__VA_ARGS__) -#define svshllb_n_u64(...) __builtin_sve_svshllb_n_u64(__VA_ARGS__) -#define svshllb_n_u16(...) __builtin_sve_svshllb_n_u16(__VA_ARGS__) -#define svshllt_n_s32(...) __builtin_sve_svshllt_n_s32(__VA_ARGS__) -#define svshllt_n_s64(...) __builtin_sve_svshllt_n_s64(__VA_ARGS__) -#define svshllt_n_s16(...) __builtin_sve_svshllt_n_s16(__VA_ARGS__) -#define svshllt_n_u32(...) __builtin_sve_svshllt_n_u32(__VA_ARGS__) -#define svshllt_n_u64(...) __builtin_sve_svshllt_n_u64(__VA_ARGS__) -#define svshllt_n_u16(...) __builtin_sve_svshllt_n_u16(__VA_ARGS__) -#define svshrnb_n_u32(...) __builtin_sve_svshrnb_n_u32(__VA_ARGS__) -#define svshrnb_n_u64(...) __builtin_sve_svshrnb_n_u64(__VA_ARGS__) -#define svshrnb_n_u16(...) __builtin_sve_svshrnb_n_u16(__VA_ARGS__) -#define svshrnb_n_s32(...) __builtin_sve_svshrnb_n_s32(__VA_ARGS__) -#define svshrnb_n_s64(...) __builtin_sve_svshrnb_n_s64(__VA_ARGS__) -#define svshrnb_n_s16(...) __builtin_sve_svshrnb_n_s16(__VA_ARGS__) -#define svshrnt_n_u32(...) __builtin_sve_svshrnt_n_u32(__VA_ARGS__) -#define svshrnt_n_u64(...) __builtin_sve_svshrnt_n_u64(__VA_ARGS__) -#define svshrnt_n_u16(...) __builtin_sve_svshrnt_n_u16(__VA_ARGS__) -#define svshrnt_n_s32(...) __builtin_sve_svshrnt_n_s32(__VA_ARGS__) -#define svshrnt_n_s64(...) __builtin_sve_svshrnt_n_s64(__VA_ARGS__) -#define svshrnt_n_s16(...) __builtin_sve_svshrnt_n_s16(__VA_ARGS__) -#define svsli_n_u8(...) __builtin_sve_svsli_n_u8(__VA_ARGS__) -#define svsli_n_u32(...) __builtin_sve_svsli_n_u32(__VA_ARGS__) -#define svsli_n_u64(...) __builtin_sve_svsli_n_u64(__VA_ARGS__) -#define svsli_n_u16(...) __builtin_sve_svsli_n_u16(__VA_ARGS__) -#define svsli_n_s8(...) __builtin_sve_svsli_n_s8(__VA_ARGS__) -#define svsli_n_s32(...) __builtin_sve_svsli_n_s32(__VA_ARGS__) -#define svsli_n_s64(...) __builtin_sve_svsli_n_s64(__VA_ARGS__) -#define svsli_n_s16(...) __builtin_sve_svsli_n_s16(__VA_ARGS__) -#define svsqadd_n_u8_m(...) __builtin_sve_svsqadd_n_u8_m(__VA_ARGS__) -#define svsqadd_n_u32_m(...) __builtin_sve_svsqadd_n_u32_m(__VA_ARGS__) -#define svsqadd_n_u64_m(...) __builtin_sve_svsqadd_n_u64_m(__VA_ARGS__) -#define svsqadd_n_u16_m(...) __builtin_sve_svsqadd_n_u16_m(__VA_ARGS__) -#define svsqadd_n_u8_x(...) __builtin_sve_svsqadd_n_u8_x(__VA_ARGS__) -#define svsqadd_n_u32_x(...) __builtin_sve_svsqadd_n_u32_x(__VA_ARGS__) -#define svsqadd_n_u64_x(...) __builtin_sve_svsqadd_n_u64_x(__VA_ARGS__) -#define svsqadd_n_u16_x(...) __builtin_sve_svsqadd_n_u16_x(__VA_ARGS__) -#define svsqadd_n_u8_z(...) __builtin_sve_svsqadd_n_u8_z(__VA_ARGS__) -#define svsqadd_n_u32_z(...) __builtin_sve_svsqadd_n_u32_z(__VA_ARGS__) -#define svsqadd_n_u64_z(...) __builtin_sve_svsqadd_n_u64_z(__VA_ARGS__) -#define svsqadd_n_u16_z(...) __builtin_sve_svsqadd_n_u16_z(__VA_ARGS__) -#define svsqadd_u8_m(...) __builtin_sve_svsqadd_u8_m(__VA_ARGS__) -#define svsqadd_u32_m(...) __builtin_sve_svsqadd_u32_m(__VA_ARGS__) -#define svsqadd_u64_m(...) __builtin_sve_svsqadd_u64_m(__VA_ARGS__) -#define svsqadd_u16_m(...) __builtin_sve_svsqadd_u16_m(__VA_ARGS__) -#define svsqadd_u8_x(...) __builtin_sve_svsqadd_u8_x(__VA_ARGS__) -#define svsqadd_u32_x(...) __builtin_sve_svsqadd_u32_x(__VA_ARGS__) -#define svsqadd_u64_x(...) __builtin_sve_svsqadd_u64_x(__VA_ARGS__) -#define svsqadd_u16_x(...) __builtin_sve_svsqadd_u16_x(__VA_ARGS__) -#define svsqadd_u8_z(...) __builtin_sve_svsqadd_u8_z(__VA_ARGS__) -#define svsqadd_u32_z(...) __builtin_sve_svsqadd_u32_z(__VA_ARGS__) -#define svsqadd_u64_z(...) __builtin_sve_svsqadd_u64_z(__VA_ARGS__) -#define svsqadd_u16_z(...) __builtin_sve_svsqadd_u16_z(__VA_ARGS__) -#define svsra_n_s8(...) __builtin_sve_svsra_n_s8(__VA_ARGS__) -#define svsra_n_s32(...) __builtin_sve_svsra_n_s32(__VA_ARGS__) -#define svsra_n_s64(...) __builtin_sve_svsra_n_s64(__VA_ARGS__) -#define svsra_n_s16(...) __builtin_sve_svsra_n_s16(__VA_ARGS__) -#define svsra_n_u8(...) __builtin_sve_svsra_n_u8(__VA_ARGS__) -#define svsra_n_u32(...) __builtin_sve_svsra_n_u32(__VA_ARGS__) -#define svsra_n_u64(...) __builtin_sve_svsra_n_u64(__VA_ARGS__) -#define svsra_n_u16(...) __builtin_sve_svsra_n_u16(__VA_ARGS__) -#define svsri_n_u8(...) __builtin_sve_svsri_n_u8(__VA_ARGS__) -#define svsri_n_u32(...) __builtin_sve_svsri_n_u32(__VA_ARGS__) -#define svsri_n_u64(...) __builtin_sve_svsri_n_u64(__VA_ARGS__) -#define svsri_n_u16(...) __builtin_sve_svsri_n_u16(__VA_ARGS__) -#define svsri_n_s8(...) __builtin_sve_svsri_n_s8(__VA_ARGS__) -#define svsri_n_s32(...) __builtin_sve_svsri_n_s32(__VA_ARGS__) -#define svsri_n_s64(...) __builtin_sve_svsri_n_s64(__VA_ARGS__) -#define svsri_n_s16(...) __builtin_sve_svsri_n_s16(__VA_ARGS__) -#define svstnt1_scatter_u32base_index_u32(...) __builtin_sve_svstnt1_scatter_u32base_index_u32(__VA_ARGS__) -#define svstnt1_scatter_u64base_index_u64(...) __builtin_sve_svstnt1_scatter_u64base_index_u64(__VA_ARGS__) -#define svstnt1_scatter_u64base_index_f64(...) __builtin_sve_svstnt1_scatter_u64base_index_f64(__VA_ARGS__) -#define svstnt1_scatter_u32base_index_f32(...) __builtin_sve_svstnt1_scatter_u32base_index_f32(__VA_ARGS__) -#define svstnt1_scatter_u32base_index_s32(...) __builtin_sve_svstnt1_scatter_u32base_index_s32(__VA_ARGS__) -#define svstnt1_scatter_u64base_index_s64(...) __builtin_sve_svstnt1_scatter_u64base_index_s64(__VA_ARGS__) -#define svstnt1_scatter_u32base_offset_u32(...) __builtin_sve_svstnt1_scatter_u32base_offset_u32(__VA_ARGS__) -#define svstnt1_scatter_u64base_offset_u64(...) __builtin_sve_svstnt1_scatter_u64base_offset_u64(__VA_ARGS__) -#define svstnt1_scatter_u64base_offset_f64(...) __builtin_sve_svstnt1_scatter_u64base_offset_f64(__VA_ARGS__) -#define svstnt1_scatter_u32base_offset_f32(...) __builtin_sve_svstnt1_scatter_u32base_offset_f32(__VA_ARGS__) -#define svstnt1_scatter_u32base_offset_s32(...) __builtin_sve_svstnt1_scatter_u32base_offset_s32(__VA_ARGS__) -#define svstnt1_scatter_u64base_offset_s64(...) __builtin_sve_svstnt1_scatter_u64base_offset_s64(__VA_ARGS__) -#define svstnt1_scatter_u32base_u32(...) __builtin_sve_svstnt1_scatter_u32base_u32(__VA_ARGS__) -#define svstnt1_scatter_u64base_u64(...) __builtin_sve_svstnt1_scatter_u64base_u64(__VA_ARGS__) -#define svstnt1_scatter_u64base_f64(...) __builtin_sve_svstnt1_scatter_u64base_f64(__VA_ARGS__) -#define svstnt1_scatter_u32base_f32(...) __builtin_sve_svstnt1_scatter_u32base_f32(__VA_ARGS__) -#define svstnt1_scatter_u32base_s32(...) __builtin_sve_svstnt1_scatter_u32base_s32(__VA_ARGS__) -#define svstnt1_scatter_u64base_s64(...) __builtin_sve_svstnt1_scatter_u64base_s64(__VA_ARGS__) -#define svstnt1_scatter_s64index_u64(...) __builtin_sve_svstnt1_scatter_s64index_u64(__VA_ARGS__) -#define svstnt1_scatter_s64index_f64(...) __builtin_sve_svstnt1_scatter_s64index_f64(__VA_ARGS__) -#define svstnt1_scatter_s64index_s64(...) __builtin_sve_svstnt1_scatter_s64index_s64(__VA_ARGS__) -#define svstnt1_scatter_u64index_u64(...) __builtin_sve_svstnt1_scatter_u64index_u64(__VA_ARGS__) -#define svstnt1_scatter_u64index_f64(...) __builtin_sve_svstnt1_scatter_u64index_f64(__VA_ARGS__) -#define svstnt1_scatter_u64index_s64(...) __builtin_sve_svstnt1_scatter_u64index_s64(__VA_ARGS__) -#define svstnt1_scatter_u32offset_u32(...) __builtin_sve_svstnt1_scatter_u32offset_u32(__VA_ARGS__) -#define svstnt1_scatter_u32offset_f32(...) __builtin_sve_svstnt1_scatter_u32offset_f32(__VA_ARGS__) -#define svstnt1_scatter_u32offset_s32(...) __builtin_sve_svstnt1_scatter_u32offset_s32(__VA_ARGS__) -#define svstnt1_scatter_s64offset_u64(...) __builtin_sve_svstnt1_scatter_s64offset_u64(__VA_ARGS__) -#define svstnt1_scatter_s64offset_f64(...) __builtin_sve_svstnt1_scatter_s64offset_f64(__VA_ARGS__) -#define svstnt1_scatter_s64offset_s64(...) __builtin_sve_svstnt1_scatter_s64offset_s64(__VA_ARGS__) -#define svstnt1_scatter_u64offset_u64(...) __builtin_sve_svstnt1_scatter_u64offset_u64(__VA_ARGS__) -#define svstnt1_scatter_u64offset_f64(...) __builtin_sve_svstnt1_scatter_u64offset_f64(__VA_ARGS__) -#define svstnt1_scatter_u64offset_s64(...) __builtin_sve_svstnt1_scatter_u64offset_s64(__VA_ARGS__) -#define svstnt1b_scatter_u32base_offset_u32(...) __builtin_sve_svstnt1b_scatter_u32base_offset_u32(__VA_ARGS__) -#define svstnt1b_scatter_u64base_offset_u64(...) __builtin_sve_svstnt1b_scatter_u64base_offset_u64(__VA_ARGS__) -#define svstnt1b_scatter_u32base_offset_s32(...) __builtin_sve_svstnt1b_scatter_u32base_offset_s32(__VA_ARGS__) -#define svstnt1b_scatter_u64base_offset_s64(...) __builtin_sve_svstnt1b_scatter_u64base_offset_s64(__VA_ARGS__) -#define svstnt1b_scatter_u32base_u32(...) __builtin_sve_svstnt1b_scatter_u32base_u32(__VA_ARGS__) -#define svstnt1b_scatter_u64base_u64(...) __builtin_sve_svstnt1b_scatter_u64base_u64(__VA_ARGS__) -#define svstnt1b_scatter_u32base_s32(...) __builtin_sve_svstnt1b_scatter_u32base_s32(__VA_ARGS__) -#define svstnt1b_scatter_u64base_s64(...) __builtin_sve_svstnt1b_scatter_u64base_s64(__VA_ARGS__) -#define svstnt1b_scatter_u32offset_s32(...) __builtin_sve_svstnt1b_scatter_u32offset_s32(__VA_ARGS__) -#define svstnt1b_scatter_u32offset_u32(...) __builtin_sve_svstnt1b_scatter_u32offset_u32(__VA_ARGS__) -#define svstnt1b_scatter_s64offset_s64(...) __builtin_sve_svstnt1b_scatter_s64offset_s64(__VA_ARGS__) -#define svstnt1b_scatter_s64offset_u64(...) __builtin_sve_svstnt1b_scatter_s64offset_u64(__VA_ARGS__) -#define svstnt1b_scatter_u64offset_s64(...) __builtin_sve_svstnt1b_scatter_u64offset_s64(__VA_ARGS__) -#define svstnt1b_scatter_u64offset_u64(...) __builtin_sve_svstnt1b_scatter_u64offset_u64(__VA_ARGS__) -#define svstnt1h_scatter_u32base_index_u32(...) __builtin_sve_svstnt1h_scatter_u32base_index_u32(__VA_ARGS__) -#define svstnt1h_scatter_u64base_index_u64(...) __builtin_sve_svstnt1h_scatter_u64base_index_u64(__VA_ARGS__) -#define svstnt1h_scatter_u32base_index_s32(...) __builtin_sve_svstnt1h_scatter_u32base_index_s32(__VA_ARGS__) -#define svstnt1h_scatter_u64base_index_s64(...) __builtin_sve_svstnt1h_scatter_u64base_index_s64(__VA_ARGS__) -#define svstnt1h_scatter_u32base_offset_u32(...) __builtin_sve_svstnt1h_scatter_u32base_offset_u32(__VA_ARGS__) -#define svstnt1h_scatter_u64base_offset_u64(...) __builtin_sve_svstnt1h_scatter_u64base_offset_u64(__VA_ARGS__) -#define svstnt1h_scatter_u32base_offset_s32(...) __builtin_sve_svstnt1h_scatter_u32base_offset_s32(__VA_ARGS__) -#define svstnt1h_scatter_u64base_offset_s64(...) __builtin_sve_svstnt1h_scatter_u64base_offset_s64(__VA_ARGS__) -#define svstnt1h_scatter_u32base_u32(...) __builtin_sve_svstnt1h_scatter_u32base_u32(__VA_ARGS__) -#define svstnt1h_scatter_u64base_u64(...) __builtin_sve_svstnt1h_scatter_u64base_u64(__VA_ARGS__) -#define svstnt1h_scatter_u32base_s32(...) __builtin_sve_svstnt1h_scatter_u32base_s32(__VA_ARGS__) -#define svstnt1h_scatter_u64base_s64(...) __builtin_sve_svstnt1h_scatter_u64base_s64(__VA_ARGS__) -#define svstnt1h_scatter_s64index_s64(...) __builtin_sve_svstnt1h_scatter_s64index_s64(__VA_ARGS__) -#define svstnt1h_scatter_s64index_u64(...) __builtin_sve_svstnt1h_scatter_s64index_u64(__VA_ARGS__) -#define svstnt1h_scatter_u64index_s64(...) __builtin_sve_svstnt1h_scatter_u64index_s64(__VA_ARGS__) -#define svstnt1h_scatter_u64index_u64(...) __builtin_sve_svstnt1h_scatter_u64index_u64(__VA_ARGS__) -#define svstnt1h_scatter_u32offset_s32(...) __builtin_sve_svstnt1h_scatter_u32offset_s32(__VA_ARGS__) -#define svstnt1h_scatter_u32offset_u32(...) __builtin_sve_svstnt1h_scatter_u32offset_u32(__VA_ARGS__) -#define svstnt1h_scatter_s64offset_s64(...) __builtin_sve_svstnt1h_scatter_s64offset_s64(__VA_ARGS__) -#define svstnt1h_scatter_s64offset_u64(...) __builtin_sve_svstnt1h_scatter_s64offset_u64(__VA_ARGS__) -#define svstnt1h_scatter_u64offset_s64(...) __builtin_sve_svstnt1h_scatter_u64offset_s64(__VA_ARGS__) -#define svstnt1h_scatter_u64offset_u64(...) __builtin_sve_svstnt1h_scatter_u64offset_u64(__VA_ARGS__) -#define svstnt1w_scatter_u64base_index_u64(...) __builtin_sve_svstnt1w_scatter_u64base_index_u64(__VA_ARGS__) -#define svstnt1w_scatter_u64base_index_s64(...) __builtin_sve_svstnt1w_scatter_u64base_index_s64(__VA_ARGS__) -#define svstnt1w_scatter_u64base_offset_u64(...) __builtin_sve_svstnt1w_scatter_u64base_offset_u64(__VA_ARGS__) -#define svstnt1w_scatter_u64base_offset_s64(...) __builtin_sve_svstnt1w_scatter_u64base_offset_s64(__VA_ARGS__) -#define svstnt1w_scatter_u64base_u64(...) __builtin_sve_svstnt1w_scatter_u64base_u64(__VA_ARGS__) -#define svstnt1w_scatter_u64base_s64(...) __builtin_sve_svstnt1w_scatter_u64base_s64(__VA_ARGS__) -#define svstnt1w_scatter_s64index_s64(...) __builtin_sve_svstnt1w_scatter_s64index_s64(__VA_ARGS__) -#define svstnt1w_scatter_s64index_u64(...) __builtin_sve_svstnt1w_scatter_s64index_u64(__VA_ARGS__) -#define svstnt1w_scatter_u64index_s64(...) __builtin_sve_svstnt1w_scatter_u64index_s64(__VA_ARGS__) -#define svstnt1w_scatter_u64index_u64(...) __builtin_sve_svstnt1w_scatter_u64index_u64(__VA_ARGS__) -#define svstnt1w_scatter_s64offset_s64(...) __builtin_sve_svstnt1w_scatter_s64offset_s64(__VA_ARGS__) -#define svstnt1w_scatter_s64offset_u64(...) __builtin_sve_svstnt1w_scatter_s64offset_u64(__VA_ARGS__) -#define svstnt1w_scatter_u64offset_s64(...) __builtin_sve_svstnt1w_scatter_u64offset_s64(__VA_ARGS__) -#define svstnt1w_scatter_u64offset_u64(...) __builtin_sve_svstnt1w_scatter_u64offset_u64(__VA_ARGS__) -#define svsubhnb_n_u32(...) __builtin_sve_svsubhnb_n_u32(__VA_ARGS__) -#define svsubhnb_n_u64(...) __builtin_sve_svsubhnb_n_u64(__VA_ARGS__) -#define svsubhnb_n_u16(...) __builtin_sve_svsubhnb_n_u16(__VA_ARGS__) -#define svsubhnb_n_s32(...) __builtin_sve_svsubhnb_n_s32(__VA_ARGS__) -#define svsubhnb_n_s64(...) __builtin_sve_svsubhnb_n_s64(__VA_ARGS__) -#define svsubhnb_n_s16(...) __builtin_sve_svsubhnb_n_s16(__VA_ARGS__) -#define svsubhnb_u32(...) __builtin_sve_svsubhnb_u32(__VA_ARGS__) -#define svsubhnb_u64(...) __builtin_sve_svsubhnb_u64(__VA_ARGS__) -#define svsubhnb_u16(...) __builtin_sve_svsubhnb_u16(__VA_ARGS__) -#define svsubhnb_s32(...) __builtin_sve_svsubhnb_s32(__VA_ARGS__) -#define svsubhnb_s64(...) __builtin_sve_svsubhnb_s64(__VA_ARGS__) -#define svsubhnb_s16(...) __builtin_sve_svsubhnb_s16(__VA_ARGS__) -#define svsubhnt_n_u32(...) __builtin_sve_svsubhnt_n_u32(__VA_ARGS__) -#define svsubhnt_n_u64(...) __builtin_sve_svsubhnt_n_u64(__VA_ARGS__) -#define svsubhnt_n_u16(...) __builtin_sve_svsubhnt_n_u16(__VA_ARGS__) -#define svsubhnt_n_s32(...) __builtin_sve_svsubhnt_n_s32(__VA_ARGS__) -#define svsubhnt_n_s64(...) __builtin_sve_svsubhnt_n_s64(__VA_ARGS__) -#define svsubhnt_n_s16(...) __builtin_sve_svsubhnt_n_s16(__VA_ARGS__) -#define svsubhnt_u32(...) __builtin_sve_svsubhnt_u32(__VA_ARGS__) -#define svsubhnt_u64(...) __builtin_sve_svsubhnt_u64(__VA_ARGS__) -#define svsubhnt_u16(...) __builtin_sve_svsubhnt_u16(__VA_ARGS__) -#define svsubhnt_s32(...) __builtin_sve_svsubhnt_s32(__VA_ARGS__) -#define svsubhnt_s64(...) __builtin_sve_svsubhnt_s64(__VA_ARGS__) -#define svsubhnt_s16(...) __builtin_sve_svsubhnt_s16(__VA_ARGS__) -#define svsublb_n_s32(...) __builtin_sve_svsublb_n_s32(__VA_ARGS__) -#define svsublb_n_s64(...) __builtin_sve_svsublb_n_s64(__VA_ARGS__) -#define svsublb_n_s16(...) __builtin_sve_svsublb_n_s16(__VA_ARGS__) -#define svsublb_n_u32(...) __builtin_sve_svsublb_n_u32(__VA_ARGS__) -#define svsublb_n_u64(...) __builtin_sve_svsublb_n_u64(__VA_ARGS__) -#define svsublb_n_u16(...) __builtin_sve_svsublb_n_u16(__VA_ARGS__) -#define svsublb_s32(...) __builtin_sve_svsublb_s32(__VA_ARGS__) -#define svsublb_s64(...) __builtin_sve_svsublb_s64(__VA_ARGS__) -#define svsublb_s16(...) __builtin_sve_svsublb_s16(__VA_ARGS__) -#define svsublb_u32(...) __builtin_sve_svsublb_u32(__VA_ARGS__) -#define svsublb_u64(...) __builtin_sve_svsublb_u64(__VA_ARGS__) -#define svsublb_u16(...) __builtin_sve_svsublb_u16(__VA_ARGS__) -#define svsublbt_n_s32(...) __builtin_sve_svsublbt_n_s32(__VA_ARGS__) -#define svsublbt_n_s64(...) __builtin_sve_svsublbt_n_s64(__VA_ARGS__) -#define svsublbt_n_s16(...) __builtin_sve_svsublbt_n_s16(__VA_ARGS__) -#define svsublbt_s32(...) __builtin_sve_svsublbt_s32(__VA_ARGS__) -#define svsublbt_s64(...) __builtin_sve_svsublbt_s64(__VA_ARGS__) -#define svsublbt_s16(...) __builtin_sve_svsublbt_s16(__VA_ARGS__) -#define svsublt_n_s32(...) __builtin_sve_svsublt_n_s32(__VA_ARGS__) -#define svsublt_n_s64(...) __builtin_sve_svsublt_n_s64(__VA_ARGS__) -#define svsublt_n_s16(...) __builtin_sve_svsublt_n_s16(__VA_ARGS__) -#define svsublt_n_u32(...) __builtin_sve_svsublt_n_u32(__VA_ARGS__) -#define svsublt_n_u64(...) __builtin_sve_svsublt_n_u64(__VA_ARGS__) -#define svsublt_n_u16(...) __builtin_sve_svsublt_n_u16(__VA_ARGS__) -#define svsublt_s32(...) __builtin_sve_svsublt_s32(__VA_ARGS__) -#define svsublt_s64(...) __builtin_sve_svsublt_s64(__VA_ARGS__) -#define svsublt_s16(...) __builtin_sve_svsublt_s16(__VA_ARGS__) -#define svsublt_u32(...) __builtin_sve_svsublt_u32(__VA_ARGS__) -#define svsublt_u64(...) __builtin_sve_svsublt_u64(__VA_ARGS__) -#define svsublt_u16(...) __builtin_sve_svsublt_u16(__VA_ARGS__) -#define svsubltb_n_s32(...) __builtin_sve_svsubltb_n_s32(__VA_ARGS__) -#define svsubltb_n_s64(...) __builtin_sve_svsubltb_n_s64(__VA_ARGS__) -#define svsubltb_n_s16(...) __builtin_sve_svsubltb_n_s16(__VA_ARGS__) -#define svsubltb_s32(...) __builtin_sve_svsubltb_s32(__VA_ARGS__) -#define svsubltb_s64(...) __builtin_sve_svsubltb_s64(__VA_ARGS__) -#define svsubltb_s16(...) __builtin_sve_svsubltb_s16(__VA_ARGS__) -#define svsubwb_n_s32(...) __builtin_sve_svsubwb_n_s32(__VA_ARGS__) -#define svsubwb_n_s64(...) __builtin_sve_svsubwb_n_s64(__VA_ARGS__) -#define svsubwb_n_s16(...) __builtin_sve_svsubwb_n_s16(__VA_ARGS__) -#define svsubwb_n_u32(...) __builtin_sve_svsubwb_n_u32(__VA_ARGS__) -#define svsubwb_n_u64(...) __builtin_sve_svsubwb_n_u64(__VA_ARGS__) -#define svsubwb_n_u16(...) __builtin_sve_svsubwb_n_u16(__VA_ARGS__) -#define svsubwb_s32(...) __builtin_sve_svsubwb_s32(__VA_ARGS__) -#define svsubwb_s64(...) __builtin_sve_svsubwb_s64(__VA_ARGS__) -#define svsubwb_s16(...) __builtin_sve_svsubwb_s16(__VA_ARGS__) -#define svsubwb_u32(...) __builtin_sve_svsubwb_u32(__VA_ARGS__) -#define svsubwb_u64(...) __builtin_sve_svsubwb_u64(__VA_ARGS__) -#define svsubwb_u16(...) __builtin_sve_svsubwb_u16(__VA_ARGS__) -#define svsubwt_n_s32(...) __builtin_sve_svsubwt_n_s32(__VA_ARGS__) -#define svsubwt_n_s64(...) __builtin_sve_svsubwt_n_s64(__VA_ARGS__) -#define svsubwt_n_s16(...) __builtin_sve_svsubwt_n_s16(__VA_ARGS__) -#define svsubwt_n_u32(...) __builtin_sve_svsubwt_n_u32(__VA_ARGS__) -#define svsubwt_n_u64(...) __builtin_sve_svsubwt_n_u64(__VA_ARGS__) -#define svsubwt_n_u16(...) __builtin_sve_svsubwt_n_u16(__VA_ARGS__) -#define svsubwt_s32(...) __builtin_sve_svsubwt_s32(__VA_ARGS__) -#define svsubwt_s64(...) __builtin_sve_svsubwt_s64(__VA_ARGS__) -#define svsubwt_s16(...) __builtin_sve_svsubwt_s16(__VA_ARGS__) -#define svsubwt_u32(...) __builtin_sve_svsubwt_u32(__VA_ARGS__) -#define svsubwt_u64(...) __builtin_sve_svsubwt_u64(__VA_ARGS__) -#define svsubwt_u16(...) __builtin_sve_svsubwt_u16(__VA_ARGS__) -#define svtbl2_u8(...) __builtin_sve_svtbl2_u8(__VA_ARGS__) -#define svtbl2_u32(...) __builtin_sve_svtbl2_u32(__VA_ARGS__) -#define svtbl2_u64(...) __builtin_sve_svtbl2_u64(__VA_ARGS__) -#define svtbl2_u16(...) __builtin_sve_svtbl2_u16(__VA_ARGS__) -#define svtbl2_s8(...) __builtin_sve_svtbl2_s8(__VA_ARGS__) -#define svtbl2_f64(...) __builtin_sve_svtbl2_f64(__VA_ARGS__) -#define svtbl2_f32(...) __builtin_sve_svtbl2_f32(__VA_ARGS__) -#define svtbl2_f16(...) __builtin_sve_svtbl2_f16(__VA_ARGS__) -#define svtbl2_s32(...) __builtin_sve_svtbl2_s32(__VA_ARGS__) -#define svtbl2_s64(...) __builtin_sve_svtbl2_s64(__VA_ARGS__) -#define svtbl2_s16(...) __builtin_sve_svtbl2_s16(__VA_ARGS__) -#define svtbx_u8(...) __builtin_sve_svtbx_u8(__VA_ARGS__) -#define svtbx_u32(...) __builtin_sve_svtbx_u32(__VA_ARGS__) -#define svtbx_u64(...) __builtin_sve_svtbx_u64(__VA_ARGS__) -#define svtbx_u16(...) __builtin_sve_svtbx_u16(__VA_ARGS__) -#define svtbx_s8(...) __builtin_sve_svtbx_s8(__VA_ARGS__) -#define svtbx_f64(...) __builtin_sve_svtbx_f64(__VA_ARGS__) -#define svtbx_f32(...) __builtin_sve_svtbx_f32(__VA_ARGS__) -#define svtbx_f16(...) __builtin_sve_svtbx_f16(__VA_ARGS__) -#define svtbx_s32(...) __builtin_sve_svtbx_s32(__VA_ARGS__) -#define svtbx_s64(...) __builtin_sve_svtbx_s64(__VA_ARGS__) -#define svtbx_s16(...) __builtin_sve_svtbx_s16(__VA_ARGS__) -#define svuqadd_n_s8_m(...) __builtin_sve_svuqadd_n_s8_m(__VA_ARGS__) -#define svuqadd_n_s32_m(...) __builtin_sve_svuqadd_n_s32_m(__VA_ARGS__) -#define svuqadd_n_s64_m(...) __builtin_sve_svuqadd_n_s64_m(__VA_ARGS__) -#define svuqadd_n_s16_m(...) __builtin_sve_svuqadd_n_s16_m(__VA_ARGS__) -#define svuqadd_n_s8_x(...) __builtin_sve_svuqadd_n_s8_x(__VA_ARGS__) -#define svuqadd_n_s32_x(...) __builtin_sve_svuqadd_n_s32_x(__VA_ARGS__) -#define svuqadd_n_s64_x(...) __builtin_sve_svuqadd_n_s64_x(__VA_ARGS__) -#define svuqadd_n_s16_x(...) __builtin_sve_svuqadd_n_s16_x(__VA_ARGS__) -#define svuqadd_n_s8_z(...) __builtin_sve_svuqadd_n_s8_z(__VA_ARGS__) -#define svuqadd_n_s32_z(...) __builtin_sve_svuqadd_n_s32_z(__VA_ARGS__) -#define svuqadd_n_s64_z(...) __builtin_sve_svuqadd_n_s64_z(__VA_ARGS__) -#define svuqadd_n_s16_z(...) __builtin_sve_svuqadd_n_s16_z(__VA_ARGS__) -#define svuqadd_s8_m(...) __builtin_sve_svuqadd_s8_m(__VA_ARGS__) -#define svuqadd_s32_m(...) __builtin_sve_svuqadd_s32_m(__VA_ARGS__) -#define svuqadd_s64_m(...) __builtin_sve_svuqadd_s64_m(__VA_ARGS__) -#define svuqadd_s16_m(...) __builtin_sve_svuqadd_s16_m(__VA_ARGS__) -#define svuqadd_s8_x(...) __builtin_sve_svuqadd_s8_x(__VA_ARGS__) -#define svuqadd_s32_x(...) __builtin_sve_svuqadd_s32_x(__VA_ARGS__) -#define svuqadd_s64_x(...) __builtin_sve_svuqadd_s64_x(__VA_ARGS__) -#define svuqadd_s16_x(...) __builtin_sve_svuqadd_s16_x(__VA_ARGS__) -#define svuqadd_s8_z(...) __builtin_sve_svuqadd_s8_z(__VA_ARGS__) -#define svuqadd_s32_z(...) __builtin_sve_svuqadd_s32_z(__VA_ARGS__) -#define svuqadd_s64_z(...) __builtin_sve_svuqadd_s64_z(__VA_ARGS__) -#define svuqadd_s16_z(...) __builtin_sve_svuqadd_s16_z(__VA_ARGS__) -#define svwhilege_b8_s32(...) __builtin_sve_svwhilege_b8_s32(__VA_ARGS__) -#define svwhilege_b32_s32(...) __builtin_sve_svwhilege_b32_s32(__VA_ARGS__) -#define svwhilege_b64_s32(...) __builtin_sve_svwhilege_b64_s32(__VA_ARGS__) -#define svwhilege_b16_s32(...) __builtin_sve_svwhilege_b16_s32(__VA_ARGS__) -#define svwhilege_b8_s64(...) __builtin_sve_svwhilege_b8_s64(__VA_ARGS__) -#define svwhilege_b32_s64(...) __builtin_sve_svwhilege_b32_s64(__VA_ARGS__) -#define svwhilege_b64_s64(...) __builtin_sve_svwhilege_b64_s64(__VA_ARGS__) -#define svwhilege_b16_s64(...) __builtin_sve_svwhilege_b16_s64(__VA_ARGS__) -#define svwhilege_b8_u32(...) __builtin_sve_svwhilege_b8_u32(__VA_ARGS__) -#define svwhilege_b32_u32(...) __builtin_sve_svwhilege_b32_u32(__VA_ARGS__) -#define svwhilege_b64_u32(...) __builtin_sve_svwhilege_b64_u32(__VA_ARGS__) -#define svwhilege_b16_u32(...) __builtin_sve_svwhilege_b16_u32(__VA_ARGS__) -#define svwhilege_b8_u64(...) __builtin_sve_svwhilege_b8_u64(__VA_ARGS__) -#define svwhilege_b32_u64(...) __builtin_sve_svwhilege_b32_u64(__VA_ARGS__) -#define svwhilege_b64_u64(...) __builtin_sve_svwhilege_b64_u64(__VA_ARGS__) -#define svwhilege_b16_u64(...) __builtin_sve_svwhilege_b16_u64(__VA_ARGS__) -#define svwhilegt_b8_s32(...) __builtin_sve_svwhilegt_b8_s32(__VA_ARGS__) -#define svwhilegt_b32_s32(...) __builtin_sve_svwhilegt_b32_s32(__VA_ARGS__) -#define svwhilegt_b64_s32(...) __builtin_sve_svwhilegt_b64_s32(__VA_ARGS__) -#define svwhilegt_b16_s32(...) __builtin_sve_svwhilegt_b16_s32(__VA_ARGS__) -#define svwhilegt_b8_s64(...) __builtin_sve_svwhilegt_b8_s64(__VA_ARGS__) -#define svwhilegt_b32_s64(...) __builtin_sve_svwhilegt_b32_s64(__VA_ARGS__) -#define svwhilegt_b64_s64(...) __builtin_sve_svwhilegt_b64_s64(__VA_ARGS__) -#define svwhilegt_b16_s64(...) __builtin_sve_svwhilegt_b16_s64(__VA_ARGS__) -#define svwhilegt_b8_u32(...) __builtin_sve_svwhilegt_b8_u32(__VA_ARGS__) -#define svwhilegt_b32_u32(...) __builtin_sve_svwhilegt_b32_u32(__VA_ARGS__) -#define svwhilegt_b64_u32(...) __builtin_sve_svwhilegt_b64_u32(__VA_ARGS__) -#define svwhilegt_b16_u32(...) __builtin_sve_svwhilegt_b16_u32(__VA_ARGS__) -#define svwhilegt_b8_u64(...) __builtin_sve_svwhilegt_b8_u64(__VA_ARGS__) -#define svwhilegt_b32_u64(...) __builtin_sve_svwhilegt_b32_u64(__VA_ARGS__) -#define svwhilegt_b64_u64(...) __builtin_sve_svwhilegt_b64_u64(__VA_ARGS__) -#define svwhilegt_b16_u64(...) __builtin_sve_svwhilegt_b16_u64(__VA_ARGS__) -#define svwhilerw_u8(...) __builtin_sve_svwhilerw_u8(__VA_ARGS__) -#define svwhilerw_s8(...) __builtin_sve_svwhilerw_s8(__VA_ARGS__) -#define svwhilerw_u64(...) __builtin_sve_svwhilerw_u64(__VA_ARGS__) -#define svwhilerw_f64(...) __builtin_sve_svwhilerw_f64(__VA_ARGS__) -#define svwhilerw_s64(...) __builtin_sve_svwhilerw_s64(__VA_ARGS__) -#define svwhilerw_u16(...) __builtin_sve_svwhilerw_u16(__VA_ARGS__) -#define svwhilerw_f16(...) __builtin_sve_svwhilerw_f16(__VA_ARGS__) -#define svwhilerw_s16(...) __builtin_sve_svwhilerw_s16(__VA_ARGS__) -#define svwhilerw_u32(...) __builtin_sve_svwhilerw_u32(__VA_ARGS__) -#define svwhilerw_f32(...) __builtin_sve_svwhilerw_f32(__VA_ARGS__) -#define svwhilerw_s32(...) __builtin_sve_svwhilerw_s32(__VA_ARGS__) -#define svwhilewr_u8(...) __builtin_sve_svwhilewr_u8(__VA_ARGS__) -#define svwhilewr_s8(...) __builtin_sve_svwhilewr_s8(__VA_ARGS__) -#define svwhilewr_u64(...) __builtin_sve_svwhilewr_u64(__VA_ARGS__) -#define svwhilewr_f64(...) __builtin_sve_svwhilewr_f64(__VA_ARGS__) -#define svwhilewr_s64(...) __builtin_sve_svwhilewr_s64(__VA_ARGS__) -#define svwhilewr_u16(...) __builtin_sve_svwhilewr_u16(__VA_ARGS__) -#define svwhilewr_f16(...) __builtin_sve_svwhilewr_f16(__VA_ARGS__) -#define svwhilewr_s16(...) __builtin_sve_svwhilewr_s16(__VA_ARGS__) -#define svwhilewr_u32(...) __builtin_sve_svwhilewr_u32(__VA_ARGS__) -#define svwhilewr_f32(...) __builtin_sve_svwhilewr_f32(__VA_ARGS__) -#define svwhilewr_s32(...) __builtin_sve_svwhilewr_s32(__VA_ARGS__) -#define svxar_n_u8(...) __builtin_sve_svxar_n_u8(__VA_ARGS__) -#define svxar_n_u32(...) __builtin_sve_svxar_n_u32(__VA_ARGS__) -#define svxar_n_u64(...) __builtin_sve_svxar_n_u64(__VA_ARGS__) -#define svxar_n_u16(...) __builtin_sve_svxar_n_u16(__VA_ARGS__) -#define svxar_n_s8(...) __builtin_sve_svxar_n_s8(__VA_ARGS__) -#define svxar_n_s32(...) __builtin_sve_svxar_n_s32(__VA_ARGS__) -#define svxar_n_s64(...) __builtin_sve_svxar_n_s64(__VA_ARGS__) -#define svxar_n_s16(...) __builtin_sve_svxar_n_s16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_s8))) +svint8_t svaba_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_s32))) +svint32_t svaba_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_s64))) +svint64_t svaba_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_s16))) +svint16_t svaba_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_u8))) +svuint8_t svaba_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_u32))) +svuint32_t svaba_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_u64))) +svuint64_t svaba_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_u16))) +svuint16_t svaba_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_s8))) +svint8_t svaba_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_s32))) +svint32_t svaba_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_s64))) +svint64_t svaba_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_s16))) +svint16_t svaba_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_u8))) +svuint8_t svaba_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_u32))) +svuint32_t svaba_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_u64))) +svuint64_t svaba_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_u16))) +svuint16_t svaba_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_n_s32))) +svint32_t svabalb_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_n_s64))) +svint64_t svabalb_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_n_s16))) +svint16_t svabalb_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_n_u32))) +svuint32_t svabalb_n_u32(svuint32_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_n_u64))) +svuint64_t svabalb_n_u64(svuint64_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_n_u16))) +svuint16_t svabalb_n_u16(svuint16_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_s32))) +svint32_t svabalb_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_s64))) +svint64_t svabalb_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_s16))) +svint16_t svabalb_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_u32))) +svuint32_t svabalb_u32(svuint32_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_u64))) +svuint64_t svabalb_u64(svuint64_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_u16))) +svuint16_t svabalb_u16(svuint16_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_n_s32))) +svint32_t svabalt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_n_s64))) +svint64_t svabalt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_n_s16))) +svint16_t svabalt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_n_u32))) +svuint32_t svabalt_n_u32(svuint32_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_n_u64))) +svuint64_t svabalt_n_u64(svuint64_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_n_u16))) +svuint16_t svabalt_n_u16(svuint16_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_s32))) +svint32_t svabalt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_s64))) +svint64_t svabalt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_s16))) +svint16_t svabalt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_u32))) +svuint32_t svabalt_u32(svuint32_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_u64))) +svuint64_t svabalt_u64(svuint64_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_u16))) +svuint16_t svabalt_u16(svuint16_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_n_s32))) +svint32_t svabdlb_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_n_s64))) +svint64_t svabdlb_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_n_s16))) +svint16_t svabdlb_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_n_u32))) +svuint32_t svabdlb_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_n_u64))) +svuint64_t svabdlb_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_n_u16))) +svuint16_t svabdlb_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_s32))) +svint32_t svabdlb_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_s64))) +svint64_t svabdlb_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_s16))) +svint16_t svabdlb_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_u32))) +svuint32_t svabdlb_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_u64))) +svuint64_t svabdlb_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_u16))) +svuint16_t svabdlb_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_n_s32))) +svint32_t svabdlt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_n_s64))) +svint64_t svabdlt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_n_s16))) +svint16_t svabdlt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_n_u32))) +svuint32_t svabdlt_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_n_u64))) +svuint64_t svabdlt_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_n_u16))) +svuint16_t svabdlt_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_s32))) +svint32_t svabdlt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_s64))) +svint64_t svabdlt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_s16))) +svint16_t svabdlt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_u32))) +svuint32_t svabdlt_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_u64))) +svuint64_t svabdlt_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_u16))) +svuint16_t svabdlt_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s32_m))) +svint32_t svadalp_s32_m(svbool_t, svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s64_m))) +svint64_t svadalp_s64_m(svbool_t, svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s16_m))) +svint16_t svadalp_s16_m(svbool_t, svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s32_x))) +svint32_t svadalp_s32_x(svbool_t, svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s64_x))) +svint64_t svadalp_s64_x(svbool_t, svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s16_x))) +svint16_t svadalp_s16_x(svbool_t, svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s32_z))) +svint32_t svadalp_s32_z(svbool_t, svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s64_z))) +svint64_t svadalp_s64_z(svbool_t, svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s16_z))) +svint16_t svadalp_s16_z(svbool_t, svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u32_m))) +svuint32_t svadalp_u32_m(svbool_t, svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u64_m))) +svuint64_t svadalp_u64_m(svbool_t, svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u16_m))) +svuint16_t svadalp_u16_m(svbool_t, svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u32_x))) +svuint32_t svadalp_u32_x(svbool_t, svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u64_x))) +svuint64_t svadalp_u64_x(svbool_t, svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u16_x))) +svuint16_t svadalp_u16_x(svbool_t, svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u32_z))) +svuint32_t svadalp_u32_z(svbool_t, svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u64_z))) +svuint64_t svadalp_u64_z(svbool_t, svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u16_z))) +svuint16_t svadalp_u16_z(svbool_t, svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclb_n_u32))) +svuint32_t svadclb_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclb_n_u64))) +svuint64_t svadclb_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclb_u32))) +svuint32_t svadclb_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclb_u64))) +svuint64_t svadclb_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclt_n_u32))) +svuint32_t svadclt_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclt_n_u64))) +svuint64_t svadclt_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclt_u32))) +svuint32_t svadclt_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclt_u64))) +svuint64_t svadclt_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_n_u32))) +svuint16_t svaddhnb_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_n_u64))) +svuint32_t svaddhnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_n_u16))) +svuint8_t svaddhnb_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_n_s32))) +svint16_t svaddhnb_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_n_s64))) +svint32_t svaddhnb_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_n_s16))) +svint8_t svaddhnb_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_u32))) +svuint16_t svaddhnb_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_u64))) +svuint32_t svaddhnb_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_u16))) +svuint8_t svaddhnb_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_s32))) +svint16_t svaddhnb_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_s64))) +svint32_t svaddhnb_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_s16))) +svint8_t svaddhnb_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_n_u32))) +svuint16_t svaddhnt_n_u32(svuint16_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_n_u64))) +svuint32_t svaddhnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_n_u16))) +svuint8_t svaddhnt_n_u16(svuint8_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_n_s32))) +svint16_t svaddhnt_n_s32(svint16_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_n_s64))) +svint32_t svaddhnt_n_s64(svint32_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_n_s16))) +svint8_t svaddhnt_n_s16(svint8_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_u32))) +svuint16_t svaddhnt_u32(svuint16_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_u64))) +svuint32_t svaddhnt_u64(svuint32_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_u16))) +svuint8_t svaddhnt_u16(svuint8_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_s32))) +svint16_t svaddhnt_s32(svint16_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_s64))) +svint32_t svaddhnt_s64(svint32_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_s16))) +svint8_t svaddhnt_s16(svint8_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_n_s32))) +svint32_t svaddlb_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_n_s64))) +svint64_t svaddlb_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_n_s16))) +svint16_t svaddlb_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_n_u32))) +svuint32_t svaddlb_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_n_u64))) +svuint64_t svaddlb_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_n_u16))) +svuint16_t svaddlb_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_s32))) +svint32_t svaddlb_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_s64))) +svint64_t svaddlb_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_s16))) +svint16_t svaddlb_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_u32))) +svuint32_t svaddlb_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_u64))) +svuint64_t svaddlb_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_u16))) +svuint16_t svaddlb_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlbt_n_s32))) +svint32_t svaddlbt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlbt_n_s64))) +svint64_t svaddlbt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlbt_n_s16))) +svint16_t svaddlbt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlbt_s32))) +svint32_t svaddlbt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlbt_s64))) +svint64_t svaddlbt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlbt_s16))) +svint16_t svaddlbt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_n_s32))) +svint32_t svaddlt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_n_s64))) +svint64_t svaddlt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_n_s16))) +svint16_t svaddlt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_n_u32))) +svuint32_t svaddlt_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_n_u64))) +svuint64_t svaddlt_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_n_u16))) +svuint16_t svaddlt_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_s32))) +svint32_t svaddlt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_s64))) +svint64_t svaddlt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_s16))) +svint16_t svaddlt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_u32))) +svuint32_t svaddlt_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_u64))) +svuint64_t svaddlt_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_u16))) +svuint16_t svaddlt_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_f64_m))) +svfloat64_t svaddp_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_f32_m))) +svfloat32_t svaddp_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_f16_m))) +svfloat16_t svaddp_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_f64_x))) +svfloat64_t svaddp_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_f32_x))) +svfloat32_t svaddp_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_f16_x))) +svfloat16_t svaddp_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u8_m))) +svuint8_t svaddp_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u32_m))) +svuint32_t svaddp_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u64_m))) +svuint64_t svaddp_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u16_m))) +svuint16_t svaddp_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s8_m))) +svint8_t svaddp_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s32_m))) +svint32_t svaddp_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s64_m))) +svint64_t svaddp_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s16_m))) +svint16_t svaddp_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u8_x))) +svuint8_t svaddp_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u32_x))) +svuint32_t svaddp_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u64_x))) +svuint64_t svaddp_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u16_x))) +svuint16_t svaddp_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s8_x))) +svint8_t svaddp_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s32_x))) +svint32_t svaddp_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s64_x))) +svint64_t svaddp_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s16_x))) +svint16_t svaddp_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_n_s32))) +svint32_t svaddwb_n_s32(svint32_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_n_s64))) +svint64_t svaddwb_n_s64(svint64_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_n_s16))) +svint16_t svaddwb_n_s16(svint16_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_n_u32))) +svuint32_t svaddwb_n_u32(svuint32_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_n_u64))) +svuint64_t svaddwb_n_u64(svuint64_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_n_u16))) +svuint16_t svaddwb_n_u16(svuint16_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_s32))) +svint32_t svaddwb_s32(svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_s64))) +svint64_t svaddwb_s64(svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_s16))) +svint16_t svaddwb_s16(svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_u32))) +svuint32_t svaddwb_u32(svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_u64))) +svuint64_t svaddwb_u64(svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_u16))) +svuint16_t svaddwb_u16(svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_n_s32))) +svint32_t svaddwt_n_s32(svint32_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_n_s64))) +svint64_t svaddwt_n_s64(svint64_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_n_s16))) +svint16_t svaddwt_n_s16(svint16_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_n_u32))) +svuint32_t svaddwt_n_u32(svuint32_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_n_u64))) +svuint64_t svaddwt_n_u64(svuint64_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_n_u16))) +svuint16_t svaddwt_n_u16(svuint16_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_s32))) +svint32_t svaddwt_s32(svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_s64))) +svint64_t svaddwt_s64(svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_s16))) +svint16_t svaddwt_s16(svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_u32))) +svuint32_t svaddwt_u32(svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_u64))) +svuint64_t svaddwt_u64(svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_u16))) +svuint16_t svaddwt_u16(svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_u8))) +svuint8_t svbcax_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_u32))) +svuint32_t svbcax_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_u64))) +svuint64_t svbcax_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_u16))) +svuint16_t svbcax_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_s8))) +svint8_t svbcax_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_s32))) +svint32_t svbcax_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_s64))) +svint64_t svbcax_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_s16))) +svint16_t svbcax_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_u8))) +svuint8_t svbcax_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_u32))) +svuint32_t svbcax_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_u64))) +svuint64_t svbcax_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_u16))) +svuint16_t svbcax_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_s8))) +svint8_t svbcax_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_s32))) +svint32_t svbcax_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_s64))) +svint64_t svbcax_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_s16))) +svint16_t svbcax_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_u8))) +svuint8_t svbsl1n_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_u32))) +svuint32_t svbsl1n_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_u64))) +svuint64_t svbsl1n_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_u16))) +svuint16_t svbsl1n_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_s8))) +svint8_t svbsl1n_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_s32))) +svint32_t svbsl1n_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_s64))) +svint64_t svbsl1n_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_s16))) +svint16_t svbsl1n_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_u8))) +svuint8_t svbsl1n_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_u32))) +svuint32_t svbsl1n_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_u64))) +svuint64_t svbsl1n_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_u16))) +svuint16_t svbsl1n_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_s8))) +svint8_t svbsl1n_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_s32))) +svint32_t svbsl1n_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_s64))) +svint64_t svbsl1n_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_s16))) +svint16_t svbsl1n_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_u8))) +svuint8_t svbsl2n_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_u32))) +svuint32_t svbsl2n_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_u64))) +svuint64_t svbsl2n_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_u16))) +svuint16_t svbsl2n_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_s8))) +svint8_t svbsl2n_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_s32))) +svint32_t svbsl2n_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_s64))) +svint64_t svbsl2n_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_s16))) +svint16_t svbsl2n_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_u8))) +svuint8_t svbsl2n_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_u32))) +svuint32_t svbsl2n_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_u64))) +svuint64_t svbsl2n_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_u16))) +svuint16_t svbsl2n_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_s8))) +svint8_t svbsl2n_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_s32))) +svint32_t svbsl2n_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_s64))) +svint64_t svbsl2n_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_s16))) +svint16_t svbsl2n_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_u8))) +svuint8_t svbsl_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_u32))) +svuint32_t svbsl_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_u64))) +svuint64_t svbsl_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_u16))) +svuint16_t svbsl_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_s8))) +svint8_t svbsl_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_s32))) +svint32_t svbsl_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_s64))) +svint64_t svbsl_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_s16))) +svint16_t svbsl_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_u8))) +svuint8_t svbsl_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_u32))) +svuint32_t svbsl_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_u64))) +svuint64_t svbsl_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_u16))) +svuint16_t svbsl_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_s8))) +svint8_t svbsl_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_s32))) +svint32_t svbsl_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_s64))) +svint64_t svbsl_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_s16))) +svint16_t svbsl_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_u8))) +svuint8_t svcadd_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_u32))) +svuint32_t svcadd_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_u64))) +svuint64_t svcadd_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_u16))) +svuint16_t svcadd_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_s8))) +svint8_t svcadd_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_s32))) +svint32_t svcadd_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_s64))) +svint64_t svcadd_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_s16))) +svint16_t svcadd_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcdot_s32))) +svint32_t svcdot_s32(svint32_t, svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcdot_s64))) +svint64_t svcdot_s64(svint64_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcdot_lane_s32))) +svint32_t svcdot_lane_s32(svint32_t, svint8_t, svint8_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcdot_lane_s64))) +svint64_t svcdot_lane_s64(svint64_t, svint16_t, svint16_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_u8))) +svuint8_t svcmla_u8(svuint8_t, svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_u32))) +svuint32_t svcmla_u32(svuint32_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_u64))) +svuint64_t svcmla_u64(svuint64_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_u16))) +svuint16_t svcmla_u16(svuint16_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_s8))) +svint8_t svcmla_s8(svint8_t, svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_s32))) +svint32_t svcmla_s32(svint32_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_s64))) +svint64_t svcmla_s64(svint64_t, svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_s16))) +svint16_t svcmla_s16(svint16_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_lane_u32))) +svuint32_t svcmla_lane_u32(svuint32_t, svuint32_t, svuint32_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_lane_u16))) +svuint16_t svcmla_lane_u16(svuint16_t, svuint16_t, svuint16_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_lane_s32))) +svint32_t svcmla_lane_s32(svint32_t, svint32_t, svint32_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_lane_s16))) +svint16_t svcmla_lane_s16(svint16_t, svint16_t, svint16_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtlt_f32_f16_m))) +svfloat32_t svcvtlt_f32_f16_m(svfloat32_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtlt_f32_f16_x))) +svfloat32_t svcvtlt_f32_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtlt_f64_f32_m))) +svfloat64_t svcvtlt_f64_f32_m(svfloat64_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtlt_f64_f32_x))) +svfloat64_t svcvtlt_f64_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtnt_f16_f32_m))) +svfloat16_t svcvtnt_f16_f32_m(svfloat16_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtnt_f32_f64_m))) +svfloat32_t svcvtnt_f32_f64_m(svfloat32_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtx_f32_f64_m))) +svfloat32_t svcvtx_f32_f64_m(svfloat32_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtx_f32_f64_x))) +svfloat32_t svcvtx_f32_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtx_f32_f64_z))) +svfloat32_t svcvtx_f32_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtxnt_f32_f64_m))) +svfloat32_t svcvtxnt_f32_f64_m(svfloat32_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_u8))) +svuint8_t sveor3_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_u32))) +svuint32_t sveor3_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_u64))) +svuint64_t sveor3_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_u16))) +svuint16_t sveor3_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_s8))) +svint8_t sveor3_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_s32))) +svint32_t sveor3_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_s64))) +svint64_t sveor3_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_s16))) +svint16_t sveor3_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_u8))) +svuint8_t sveor3_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_u32))) +svuint32_t sveor3_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_u64))) +svuint64_t sveor3_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_u16))) +svuint16_t sveor3_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_s8))) +svint8_t sveor3_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_s32))) +svint32_t sveor3_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_s64))) +svint64_t sveor3_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_s16))) +svint16_t sveor3_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_u8))) +svuint8_t sveorbt_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_u32))) +svuint32_t sveorbt_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_u64))) +svuint64_t sveorbt_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_u16))) +svuint16_t sveorbt_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_s8))) +svint8_t sveorbt_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_s32))) +svint32_t sveorbt_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_s64))) +svint64_t sveorbt_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_s16))) +svint16_t sveorbt_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_u8))) +svuint8_t sveorbt_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_u32))) +svuint32_t sveorbt_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_u64))) +svuint64_t sveorbt_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_u16))) +svuint16_t sveorbt_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_s8))) +svint8_t sveorbt_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_s32))) +svint32_t sveorbt_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_s64))) +svint64_t sveorbt_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_s16))) +svint16_t sveorbt_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_u8))) +svuint8_t sveortb_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_u32))) +svuint32_t sveortb_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_u64))) +svuint64_t sveortb_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_u16))) +svuint16_t sveortb_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_s8))) +svint8_t sveortb_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_s32))) +svint32_t sveortb_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_s64))) +svint64_t sveortb_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_s16))) +svint16_t sveortb_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_u8))) +svuint8_t sveortb_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_u32))) +svuint32_t sveortb_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_u64))) +svuint64_t sveortb_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_u16))) +svuint16_t sveortb_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_s8))) +svint8_t sveortb_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_s32))) +svint32_t sveortb_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_s64))) +svint64_t sveortb_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_s16))) +svint16_t sveortb_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s8_m))) +svint8_t svhadd_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s32_m))) +svint32_t svhadd_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s64_m))) +svint64_t svhadd_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s16_m))) +svint16_t svhadd_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s8_x))) +svint8_t svhadd_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s32_x))) +svint32_t svhadd_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s64_x))) +svint64_t svhadd_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s16_x))) +svint16_t svhadd_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s8_z))) +svint8_t svhadd_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s32_z))) +svint32_t svhadd_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s64_z))) +svint64_t svhadd_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s16_z))) +svint16_t svhadd_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u8_m))) +svuint8_t svhadd_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u32_m))) +svuint32_t svhadd_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u64_m))) +svuint64_t svhadd_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u16_m))) +svuint16_t svhadd_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u8_x))) +svuint8_t svhadd_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u32_x))) +svuint32_t svhadd_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u64_x))) +svuint64_t svhadd_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u16_x))) +svuint16_t svhadd_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u8_z))) +svuint8_t svhadd_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u32_z))) +svuint32_t svhadd_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u64_z))) +svuint64_t svhadd_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u16_z))) +svuint16_t svhadd_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s8_m))) +svint8_t svhadd_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s32_m))) +svint32_t svhadd_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s64_m))) +svint64_t svhadd_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s16_m))) +svint16_t svhadd_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s8_x))) +svint8_t svhadd_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s32_x))) +svint32_t svhadd_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s64_x))) +svint64_t svhadd_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s16_x))) +svint16_t svhadd_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s8_z))) +svint8_t svhadd_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s32_z))) +svint32_t svhadd_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s64_z))) +svint64_t svhadd_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s16_z))) +svint16_t svhadd_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u8_m))) +svuint8_t svhadd_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u32_m))) +svuint32_t svhadd_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u64_m))) +svuint64_t svhadd_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u16_m))) +svuint16_t svhadd_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u8_x))) +svuint8_t svhadd_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u32_x))) +svuint32_t svhadd_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u64_x))) +svuint64_t svhadd_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u16_x))) +svuint16_t svhadd_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u8_z))) +svuint8_t svhadd_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u32_z))) +svuint32_t svhadd_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u64_z))) +svuint64_t svhadd_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u16_z))) +svuint16_t svhadd_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhistcnt_u32_z))) +svuint32_t svhistcnt_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhistcnt_u64_z))) +svuint64_t svhistcnt_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhistcnt_s32_z))) +svuint32_t svhistcnt_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhistcnt_s64_z))) +svuint64_t svhistcnt_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhistseg_u8))) +svuint8_t svhistseg_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhistseg_s8))) +svuint8_t svhistseg_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s8_m))) +svint8_t svhsub_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s32_m))) +svint32_t svhsub_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s64_m))) +svint64_t svhsub_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s16_m))) +svint16_t svhsub_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s8_x))) +svint8_t svhsub_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s32_x))) +svint32_t svhsub_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s64_x))) +svint64_t svhsub_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s16_x))) +svint16_t svhsub_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s8_z))) +svint8_t svhsub_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s32_z))) +svint32_t svhsub_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s64_z))) +svint64_t svhsub_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s16_z))) +svint16_t svhsub_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u8_m))) +svuint8_t svhsub_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u32_m))) +svuint32_t svhsub_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u64_m))) +svuint64_t svhsub_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u16_m))) +svuint16_t svhsub_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u8_x))) +svuint8_t svhsub_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u32_x))) +svuint32_t svhsub_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u64_x))) +svuint64_t svhsub_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u16_x))) +svuint16_t svhsub_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u8_z))) +svuint8_t svhsub_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u32_z))) +svuint32_t svhsub_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u64_z))) +svuint64_t svhsub_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u16_z))) +svuint16_t svhsub_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s8_m))) +svint8_t svhsub_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s32_m))) +svint32_t svhsub_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s64_m))) +svint64_t svhsub_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s16_m))) +svint16_t svhsub_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s8_x))) +svint8_t svhsub_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s32_x))) +svint32_t svhsub_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s64_x))) +svint64_t svhsub_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s16_x))) +svint16_t svhsub_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s8_z))) +svint8_t svhsub_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s32_z))) +svint32_t svhsub_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s64_z))) +svint64_t svhsub_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s16_z))) +svint16_t svhsub_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u8_m))) +svuint8_t svhsub_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u32_m))) +svuint32_t svhsub_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u64_m))) +svuint64_t svhsub_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u16_m))) +svuint16_t svhsub_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u8_x))) +svuint8_t svhsub_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u32_x))) +svuint32_t svhsub_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u64_x))) +svuint64_t svhsub_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u16_x))) +svuint16_t svhsub_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u8_z))) +svuint8_t svhsub_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u32_z))) +svuint32_t svhsub_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u64_z))) +svuint64_t svhsub_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u16_z))) +svuint16_t svhsub_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s8_m))) +svint8_t svhsubr_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s32_m))) +svint32_t svhsubr_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s64_m))) +svint64_t svhsubr_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s16_m))) +svint16_t svhsubr_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s8_x))) +svint8_t svhsubr_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s32_x))) +svint32_t svhsubr_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s64_x))) +svint64_t svhsubr_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s16_x))) +svint16_t svhsubr_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s8_z))) +svint8_t svhsubr_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s32_z))) +svint32_t svhsubr_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s64_z))) +svint64_t svhsubr_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s16_z))) +svint16_t svhsubr_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u8_m))) +svuint8_t svhsubr_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u32_m))) +svuint32_t svhsubr_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u64_m))) +svuint64_t svhsubr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u16_m))) +svuint16_t svhsubr_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u8_x))) +svuint8_t svhsubr_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u32_x))) +svuint32_t svhsubr_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u64_x))) +svuint64_t svhsubr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u16_x))) +svuint16_t svhsubr_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u8_z))) +svuint8_t svhsubr_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u32_z))) +svuint32_t svhsubr_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u64_z))) +svuint64_t svhsubr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u16_z))) +svuint16_t svhsubr_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s8_m))) +svint8_t svhsubr_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s32_m))) +svint32_t svhsubr_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s64_m))) +svint64_t svhsubr_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s16_m))) +svint16_t svhsubr_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s8_x))) +svint8_t svhsubr_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s32_x))) +svint32_t svhsubr_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s64_x))) +svint64_t svhsubr_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s16_x))) +svint16_t svhsubr_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s8_z))) +svint8_t svhsubr_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s32_z))) +svint32_t svhsubr_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s64_z))) +svint64_t svhsubr_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s16_z))) +svint16_t svhsubr_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u8_m))) +svuint8_t svhsubr_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u32_m))) +svuint32_t svhsubr_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u64_m))) +svuint64_t svhsubr_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u16_m))) +svuint16_t svhsubr_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u8_x))) +svuint8_t svhsubr_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u32_x))) +svuint32_t svhsubr_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u64_x))) +svuint64_t svhsubr_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u16_x))) +svuint16_t svhsubr_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u8_z))) +svuint8_t svhsubr_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u32_z))) +svuint32_t svhsubr_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u64_z))) +svuint64_t svhsubr_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u16_z))) +svuint16_t svhsubr_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_index_u32))) +svuint32_t svldnt1_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_index_u64))) +svuint64_t svldnt1_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_index_f64))) +svfloat64_t svldnt1_gather_u64base_index_f64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_index_f32))) +svfloat32_t svldnt1_gather_u32base_index_f32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_index_s32))) +svint32_t svldnt1_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_index_s64))) +svint64_t svldnt1_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_offset_u32))) +svuint32_t svldnt1_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_offset_u64))) +svuint64_t svldnt1_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_offset_f64))) +svfloat64_t svldnt1_gather_u64base_offset_f64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_offset_f32))) +svfloat32_t svldnt1_gather_u32base_offset_f32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_offset_s32))) +svint32_t svldnt1_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_offset_s64))) +svint64_t svldnt1_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_u32))) +svuint32_t svldnt1_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_u64))) +svuint64_t svldnt1_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_f64))) +svfloat64_t svldnt1_gather_u64base_f64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_f32))) +svfloat32_t svldnt1_gather_u32base_f32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_s32))) +svint32_t svldnt1_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_s64))) +svint64_t svldnt1_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_s64index_u64))) +svuint64_t svldnt1_gather_s64index_u64(svbool_t, uint64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_s64index_f64))) +svfloat64_t svldnt1_gather_s64index_f64(svbool_t, float64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_s64index_s64))) +svint64_t svldnt1_gather_s64index_s64(svbool_t, int64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64index_u64))) +svuint64_t svldnt1_gather_u64index_u64(svbool_t, uint64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64index_f64))) +svfloat64_t svldnt1_gather_u64index_f64(svbool_t, float64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64index_s64))) +svint64_t svldnt1_gather_u64index_s64(svbool_t, int64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32offset_u32))) +svuint32_t svldnt1_gather_u32offset_u32(svbool_t, uint32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32offset_f32))) +svfloat32_t svldnt1_gather_u32offset_f32(svbool_t, float32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32offset_s32))) +svint32_t svldnt1_gather_u32offset_s32(svbool_t, int32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_s64offset_u64))) +svuint64_t svldnt1_gather_s64offset_u64(svbool_t, uint64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_s64offset_f64))) +svfloat64_t svldnt1_gather_s64offset_f64(svbool_t, float64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_s64offset_s64))) +svint64_t svldnt1_gather_s64offset_s64(svbool_t, int64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64offset_u64))) +svuint64_t svldnt1_gather_u64offset_u64(svbool_t, uint64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64offset_f64))) +svfloat64_t svldnt1_gather_u64offset_f64(svbool_t, float64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64offset_s64))) +svint64_t svldnt1_gather_u64offset_s64(svbool_t, int64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u32base_offset_u32))) +svuint32_t svldnt1sb_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u64base_offset_u64))) +svuint64_t svldnt1sb_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u32base_offset_s32))) +svint32_t svldnt1sb_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u64base_offset_s64))) +svint64_t svldnt1sb_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u32base_u32))) +svuint32_t svldnt1sb_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u64base_u64))) +svuint64_t svldnt1sb_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u32base_s32))) +svint32_t svldnt1sb_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u64base_s64))) +svint64_t svldnt1sb_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u32offset_u32))) +svuint32_t svldnt1sb_gather_u32offset_u32(svbool_t, int8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u32offset_s32))) +svint32_t svldnt1sb_gather_u32offset_s32(svbool_t, int8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_s64offset_u64))) +svuint64_t svldnt1sb_gather_s64offset_u64(svbool_t, int8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_s64offset_s64))) +svint64_t svldnt1sb_gather_s64offset_s64(svbool_t, int8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u64offset_u64))) +svuint64_t svldnt1sb_gather_u64offset_u64(svbool_t, int8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u64offset_s64))) +svint64_t svldnt1sb_gather_u64offset_s64(svbool_t, int8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32base_index_u32))) +svuint32_t svldnt1sh_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64base_index_u64))) +svuint64_t svldnt1sh_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32base_index_s32))) +svint32_t svldnt1sh_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64base_index_s64))) +svint64_t svldnt1sh_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32base_offset_u32))) +svuint32_t svldnt1sh_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64base_offset_u64))) +svuint64_t svldnt1sh_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32base_offset_s32))) +svint32_t svldnt1sh_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64base_offset_s64))) +svint64_t svldnt1sh_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32base_u32))) +svuint32_t svldnt1sh_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64base_u64))) +svuint64_t svldnt1sh_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32base_s32))) +svint32_t svldnt1sh_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64base_s64))) +svint64_t svldnt1sh_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_s64index_u64))) +svuint64_t svldnt1sh_gather_s64index_u64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_s64index_s64))) +svint64_t svldnt1sh_gather_s64index_s64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64index_u64))) +svuint64_t svldnt1sh_gather_u64index_u64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64index_s64))) +svint64_t svldnt1sh_gather_u64index_s64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32offset_u32))) +svuint32_t svldnt1sh_gather_u32offset_u32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32offset_s32))) +svint32_t svldnt1sh_gather_u32offset_s32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_s64offset_u64))) +svuint64_t svldnt1sh_gather_s64offset_u64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_s64offset_s64))) +svint64_t svldnt1sh_gather_s64offset_s64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64offset_u64))) +svuint64_t svldnt1sh_gather_u64offset_u64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64offset_s64))) +svint64_t svldnt1sh_gather_u64offset_s64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64base_index_u64))) +svuint64_t svldnt1sw_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64base_index_s64))) +svint64_t svldnt1sw_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64base_offset_u64))) +svuint64_t svldnt1sw_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64base_offset_s64))) +svint64_t svldnt1sw_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64base_u64))) +svuint64_t svldnt1sw_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64base_s64))) +svint64_t svldnt1sw_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_s64index_u64))) +svuint64_t svldnt1sw_gather_s64index_u64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_s64index_s64))) +svint64_t svldnt1sw_gather_s64index_s64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64index_u64))) +svuint64_t svldnt1sw_gather_u64index_u64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64index_s64))) +svint64_t svldnt1sw_gather_u64index_s64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_s64offset_u64))) +svuint64_t svldnt1sw_gather_s64offset_u64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_s64offset_s64))) +svint64_t svldnt1sw_gather_s64offset_s64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64offset_u64))) +svuint64_t svldnt1sw_gather_u64offset_u64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64offset_s64))) +svint64_t svldnt1sw_gather_u64offset_s64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u32base_offset_u32))) +svuint32_t svldnt1ub_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u64base_offset_u64))) +svuint64_t svldnt1ub_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u32base_offset_s32))) +svint32_t svldnt1ub_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u64base_offset_s64))) +svint64_t svldnt1ub_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u32base_u32))) +svuint32_t svldnt1ub_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u64base_u64))) +svuint64_t svldnt1ub_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u32base_s32))) +svint32_t svldnt1ub_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u64base_s64))) +svint64_t svldnt1ub_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u32offset_u32))) +svuint32_t svldnt1ub_gather_u32offset_u32(svbool_t, uint8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u32offset_s32))) +svint32_t svldnt1ub_gather_u32offset_s32(svbool_t, uint8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_s64offset_u64))) +svuint64_t svldnt1ub_gather_s64offset_u64(svbool_t, uint8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_s64offset_s64))) +svint64_t svldnt1ub_gather_s64offset_s64(svbool_t, uint8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u64offset_u64))) +svuint64_t svldnt1ub_gather_u64offset_u64(svbool_t, uint8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u64offset_s64))) +svint64_t svldnt1ub_gather_u64offset_s64(svbool_t, uint8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32base_index_u32))) +svuint32_t svldnt1uh_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64base_index_u64))) +svuint64_t svldnt1uh_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32base_index_s32))) +svint32_t svldnt1uh_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64base_index_s64))) +svint64_t svldnt1uh_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32base_offset_u32))) +svuint32_t svldnt1uh_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64base_offset_u64))) +svuint64_t svldnt1uh_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32base_offset_s32))) +svint32_t svldnt1uh_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64base_offset_s64))) +svint64_t svldnt1uh_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32base_u32))) +svuint32_t svldnt1uh_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64base_u64))) +svuint64_t svldnt1uh_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32base_s32))) +svint32_t svldnt1uh_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64base_s64))) +svint64_t svldnt1uh_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_s64index_u64))) +svuint64_t svldnt1uh_gather_s64index_u64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_s64index_s64))) +svint64_t svldnt1uh_gather_s64index_s64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64index_u64))) +svuint64_t svldnt1uh_gather_u64index_u64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64index_s64))) +svint64_t svldnt1uh_gather_u64index_s64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32offset_u32))) +svuint32_t svldnt1uh_gather_u32offset_u32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32offset_s32))) +svint32_t svldnt1uh_gather_u32offset_s32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_s64offset_u64))) +svuint64_t svldnt1uh_gather_s64offset_u64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_s64offset_s64))) +svint64_t svldnt1uh_gather_s64offset_s64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64offset_u64))) +svuint64_t svldnt1uh_gather_u64offset_u64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64offset_s64))) +svint64_t svldnt1uh_gather_u64offset_s64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64base_index_u64))) +svuint64_t svldnt1uw_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64base_index_s64))) +svint64_t svldnt1uw_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64base_offset_u64))) +svuint64_t svldnt1uw_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64base_offset_s64))) +svint64_t svldnt1uw_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64base_u64))) +svuint64_t svldnt1uw_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64base_s64))) +svint64_t svldnt1uw_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_s64index_u64))) +svuint64_t svldnt1uw_gather_s64index_u64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_s64index_s64))) +svint64_t svldnt1uw_gather_s64index_s64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64index_u64))) +svuint64_t svldnt1uw_gather_u64index_u64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64index_s64))) +svint64_t svldnt1uw_gather_u64index_s64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_s64offset_u64))) +svuint64_t svldnt1uw_gather_s64offset_u64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_s64offset_s64))) +svint64_t svldnt1uw_gather_s64offset_s64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64offset_u64))) +svuint64_t svldnt1uw_gather_u64offset_u64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64offset_s64))) +svint64_t svldnt1uw_gather_u64offset_s64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f64_m))) +svint64_t svlogb_f64_m(svint64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f32_m))) +svint32_t svlogb_f32_m(svint32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f16_m))) +svint16_t svlogb_f16_m(svint16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f64_x))) +svint64_t svlogb_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f32_x))) +svint32_t svlogb_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f16_x))) +svint16_t svlogb_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f64_z))) +svint64_t svlogb_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f32_z))) +svint32_t svlogb_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f16_z))) +svint16_t svlogb_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmatch_u8))) +svbool_t svmatch_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmatch_u16))) +svbool_t svmatch_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmatch_s8))) +svbool_t svmatch_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmatch_s16))) +svbool_t svmatch_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmp_f64_m))) +svfloat64_t svmaxnmp_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmp_f32_m))) +svfloat32_t svmaxnmp_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmp_f16_m))) +svfloat16_t svmaxnmp_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmp_f64_x))) +svfloat64_t svmaxnmp_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmp_f32_x))) +svfloat32_t svmaxnmp_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmp_f16_x))) +svfloat16_t svmaxnmp_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_f64_m))) +svfloat64_t svmaxp_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_f32_m))) +svfloat32_t svmaxp_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_f16_m))) +svfloat16_t svmaxp_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_f64_x))) +svfloat64_t svmaxp_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_f32_x))) +svfloat32_t svmaxp_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_f16_x))) +svfloat16_t svmaxp_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s8_m))) +svint8_t svmaxp_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s32_m))) +svint32_t svmaxp_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s64_m))) +svint64_t svmaxp_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s16_m))) +svint16_t svmaxp_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s8_x))) +svint8_t svmaxp_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s32_x))) +svint32_t svmaxp_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s64_x))) +svint64_t svmaxp_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s16_x))) +svint16_t svmaxp_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u8_m))) +svuint8_t svmaxp_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u32_m))) +svuint32_t svmaxp_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u64_m))) +svuint64_t svmaxp_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u16_m))) +svuint16_t svmaxp_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u8_x))) +svuint8_t svmaxp_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u32_x))) +svuint32_t svmaxp_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u64_x))) +svuint64_t svmaxp_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u16_x))) +svuint16_t svmaxp_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmp_f64_m))) +svfloat64_t svminnmp_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmp_f32_m))) +svfloat32_t svminnmp_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmp_f16_m))) +svfloat16_t svminnmp_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmp_f64_x))) +svfloat64_t svminnmp_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmp_f32_x))) +svfloat32_t svminnmp_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmp_f16_x))) +svfloat16_t svminnmp_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_f64_m))) +svfloat64_t svminp_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_f32_m))) +svfloat32_t svminp_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_f16_m))) +svfloat16_t svminp_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_f64_x))) +svfloat64_t svminp_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_f32_x))) +svfloat32_t svminp_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_f16_x))) +svfloat16_t svminp_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s8_m))) +svint8_t svminp_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s32_m))) +svint32_t svminp_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s64_m))) +svint64_t svminp_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s16_m))) +svint16_t svminp_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s8_x))) +svint8_t svminp_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s32_x))) +svint32_t svminp_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s64_x))) +svint64_t svminp_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s16_x))) +svint16_t svminp_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u8_m))) +svuint8_t svminp_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u32_m))) +svuint32_t svminp_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u64_m))) +svuint64_t svminp_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u16_m))) +svuint16_t svminp_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u8_x))) +svuint8_t svminp_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u32_x))) +svuint32_t svminp_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u64_x))) +svuint64_t svminp_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u16_x))) +svuint16_t svminp_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_u32))) +svuint32_t svmla_lane_u32(svuint32_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_u64))) +svuint64_t svmla_lane_u64(svuint64_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_u16))) +svuint16_t svmla_lane_u16(svuint16_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_s32))) +svint32_t svmla_lane_s32(svint32_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_s64))) +svint64_t svmla_lane_s64(svint64_t, svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_s16))) +svint16_t svmla_lane_s16(svint16_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_f32))) +svfloat32_t svmlalb_n_f32(svfloat32_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_s32))) +svint32_t svmlalb_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_s64))) +svint64_t svmlalb_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_s16))) +svint16_t svmlalb_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_u32))) +svuint32_t svmlalb_n_u32(svuint32_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_u64))) +svuint64_t svmlalb_n_u64(svuint64_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_u16))) +svuint16_t svmlalb_n_u16(svuint16_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_f32))) +svfloat32_t svmlalb_f32(svfloat32_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_s32))) +svint32_t svmlalb_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_s64))) +svint64_t svmlalb_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_s16))) +svint16_t svmlalb_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_u32))) +svuint32_t svmlalb_u32(svuint32_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_u64))) +svuint64_t svmlalb_u64(svuint64_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_u16))) +svuint16_t svmlalb_u16(svuint16_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_lane_f32))) +svfloat32_t svmlalb_lane_f32(svfloat32_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_lane_s32))) +svint32_t svmlalb_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_lane_s64))) +svint64_t svmlalb_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_lane_u32))) +svuint32_t svmlalb_lane_u32(svuint32_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_lane_u64))) +svuint64_t svmlalb_lane_u64(svuint64_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_f32))) +svfloat32_t svmlalt_n_f32(svfloat32_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_s32))) +svint32_t svmlalt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_s64))) +svint64_t svmlalt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_s16))) +svint16_t svmlalt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_u32))) +svuint32_t svmlalt_n_u32(svuint32_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_u64))) +svuint64_t svmlalt_n_u64(svuint64_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_u16))) +svuint16_t svmlalt_n_u16(svuint16_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_f32))) +svfloat32_t svmlalt_f32(svfloat32_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_s32))) +svint32_t svmlalt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_s64))) +svint64_t svmlalt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_s16))) +svint16_t svmlalt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_u32))) +svuint32_t svmlalt_u32(svuint32_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_u64))) +svuint64_t svmlalt_u64(svuint64_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_u16))) +svuint16_t svmlalt_u16(svuint16_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_lane_f32))) +svfloat32_t svmlalt_lane_f32(svfloat32_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_lane_s32))) +svint32_t svmlalt_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_lane_s64))) +svint64_t svmlalt_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_lane_u32))) +svuint32_t svmlalt_lane_u32(svuint32_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_lane_u64))) +svuint64_t svmlalt_lane_u64(svuint64_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_u32))) +svuint32_t svmls_lane_u32(svuint32_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_u64))) +svuint64_t svmls_lane_u64(svuint64_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_u16))) +svuint16_t svmls_lane_u16(svuint16_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_s32))) +svint32_t svmls_lane_s32(svint32_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_s64))) +svint64_t svmls_lane_s64(svint64_t, svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_s16))) +svint16_t svmls_lane_s16(svint16_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_f32))) +svfloat32_t svmlslb_n_f32(svfloat32_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_s32))) +svint32_t svmlslb_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_s64))) +svint64_t svmlslb_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_s16))) +svint16_t svmlslb_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_u32))) +svuint32_t svmlslb_n_u32(svuint32_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_u64))) +svuint64_t svmlslb_n_u64(svuint64_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_u16))) +svuint16_t svmlslb_n_u16(svuint16_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_f32))) +svfloat32_t svmlslb_f32(svfloat32_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_s32))) +svint32_t svmlslb_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_s64))) +svint64_t svmlslb_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_s16))) +svint16_t svmlslb_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_u32))) +svuint32_t svmlslb_u32(svuint32_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_u64))) +svuint64_t svmlslb_u64(svuint64_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_u16))) +svuint16_t svmlslb_u16(svuint16_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_lane_f32))) +svfloat32_t svmlslb_lane_f32(svfloat32_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_lane_s32))) +svint32_t svmlslb_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_lane_s64))) +svint64_t svmlslb_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_lane_u32))) +svuint32_t svmlslb_lane_u32(svuint32_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_lane_u64))) +svuint64_t svmlslb_lane_u64(svuint64_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_f32))) +svfloat32_t svmlslt_n_f32(svfloat32_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_s32))) +svint32_t svmlslt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_s64))) +svint64_t svmlslt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_s16))) +svint16_t svmlslt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_u32))) +svuint32_t svmlslt_n_u32(svuint32_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_u64))) +svuint64_t svmlslt_n_u64(svuint64_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_u16))) +svuint16_t svmlslt_n_u16(svuint16_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_f32))) +svfloat32_t svmlslt_f32(svfloat32_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_s32))) +svint32_t svmlslt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_s64))) +svint64_t svmlslt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_s16))) +svint16_t svmlslt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_u32))) +svuint32_t svmlslt_u32(svuint32_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_u64))) +svuint64_t svmlslt_u64(svuint64_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_u16))) +svuint16_t svmlslt_u16(svuint16_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_lane_f32))) +svfloat32_t svmlslt_lane_f32(svfloat32_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_lane_s32))) +svint32_t svmlslt_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_lane_s64))) +svint64_t svmlslt_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_lane_u32))) +svuint32_t svmlslt_lane_u32(svuint32_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_lane_u64))) +svuint64_t svmlslt_lane_u64(svuint64_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlb_s32))) +svint32_t svmovlb_s32(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlb_s64))) +svint64_t svmovlb_s64(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlb_s16))) +svint16_t svmovlb_s16(svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlb_u32))) +svuint32_t svmovlb_u32(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlb_u64))) +svuint64_t svmovlb_u64(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlb_u16))) +svuint16_t svmovlb_u16(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlt_s32))) +svint32_t svmovlt_s32(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlt_s64))) +svint64_t svmovlt_s64(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlt_s16))) +svint16_t svmovlt_s16(svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlt_u32))) +svuint32_t svmovlt_u32(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlt_u64))) +svuint64_t svmovlt_u64(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlt_u16))) +svuint16_t svmovlt_u16(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_u32))) +svuint32_t svmul_lane_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_u64))) +svuint64_t svmul_lane_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_u16))) +svuint16_t svmul_lane_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_s32))) +svint32_t svmul_lane_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_s64))) +svint64_t svmul_lane_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_s16))) +svint16_t svmul_lane_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_n_s32))) +svint32_t svmullb_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_n_s64))) +svint64_t svmullb_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_n_s16))) +svint16_t svmullb_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_n_u32))) +svuint32_t svmullb_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_n_u64))) +svuint64_t svmullb_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_n_u16))) +svuint16_t svmullb_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_s32))) +svint32_t svmullb_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_s64))) +svint64_t svmullb_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_s16))) +svint16_t svmullb_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_u32))) +svuint32_t svmullb_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_u64))) +svuint64_t svmullb_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_u16))) +svuint16_t svmullb_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_lane_s32))) +svint32_t svmullb_lane_s32(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_lane_s64))) +svint64_t svmullb_lane_s64(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_lane_u32))) +svuint32_t svmullb_lane_u32(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_lane_u64))) +svuint64_t svmullb_lane_u64(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_n_s32))) +svint32_t svmullt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_n_s64))) +svint64_t svmullt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_n_s16))) +svint16_t svmullt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_n_u32))) +svuint32_t svmullt_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_n_u64))) +svuint64_t svmullt_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_n_u16))) +svuint16_t svmullt_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_s32))) +svint32_t svmullt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_s64))) +svint64_t svmullt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_s16))) +svint16_t svmullt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_u32))) +svuint32_t svmullt_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_u64))) +svuint64_t svmullt_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_u16))) +svuint16_t svmullt_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_lane_s32))) +svint32_t svmullt_lane_s32(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_lane_s64))) +svint64_t svmullt_lane_s64(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_lane_u32))) +svuint32_t svmullt_lane_u32(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_lane_u64))) +svuint64_t svmullt_lane_u64(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_u8))) +svuint8_t svnbsl_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_u32))) +svuint32_t svnbsl_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_u64))) +svuint64_t svnbsl_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_u16))) +svuint16_t svnbsl_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_s8))) +svint8_t svnbsl_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_s32))) +svint32_t svnbsl_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_s64))) +svint64_t svnbsl_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_s16))) +svint16_t svnbsl_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_u8))) +svuint8_t svnbsl_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_u32))) +svuint32_t svnbsl_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_u64))) +svuint64_t svnbsl_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_u16))) +svuint16_t svnbsl_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_s8))) +svint8_t svnbsl_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_s32))) +svint32_t svnbsl_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_s64))) +svint64_t svnbsl_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_s16))) +svint16_t svnbsl_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmatch_u8))) +svbool_t svnmatch_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmatch_u16))) +svbool_t svnmatch_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmatch_s8))) +svbool_t svnmatch_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmatch_s16))) +svbool_t svnmatch_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmul_n_u8))) +svuint8_t svpmul_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmul_u8))) +svuint8_t svpmul_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_n_u64))) +svuint64_t svpmullb_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_n_u16))) +svuint16_t svpmullb_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_u64))) +svuint64_t svpmullb_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_u16))) +svuint16_t svpmullb_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_pair_n_u8))) +svuint8_t svpmullb_pair_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_pair_n_u32))) +svuint32_t svpmullb_pair_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_pair_u8))) +svuint8_t svpmullb_pair_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_pair_u32))) +svuint32_t svpmullb_pair_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_n_u64))) +svuint64_t svpmullt_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_n_u16))) +svuint16_t svpmullt_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_u64))) +svuint64_t svpmullt_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_u16))) +svuint16_t svpmullt_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_pair_n_u8))) +svuint8_t svpmullt_pair_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_pair_n_u32))) +svuint32_t svpmullt_pair_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_pair_u8))) +svuint8_t svpmullt_pair_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_pair_u32))) +svuint32_t svpmullt_pair_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s8_m))) +svint8_t svqabs_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s32_m))) +svint32_t svqabs_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s64_m))) +svint64_t svqabs_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s16_m))) +svint16_t svqabs_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s8_x))) +svint8_t svqabs_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s32_x))) +svint32_t svqabs_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s64_x))) +svint64_t svqabs_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s16_x))) +svint16_t svqabs_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s8_z))) +svint8_t svqabs_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s32_z))) +svint32_t svqabs_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s64_z))) +svint64_t svqabs_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s16_z))) +svint16_t svqabs_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s8_m))) +svint8_t svqadd_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s32_m))) +svint32_t svqadd_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s64_m))) +svint64_t svqadd_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s16_m))) +svint16_t svqadd_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s8_x))) +svint8_t svqadd_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s32_x))) +svint32_t svqadd_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s64_x))) +svint64_t svqadd_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s16_x))) +svint16_t svqadd_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s8_z))) +svint8_t svqadd_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s32_z))) +svint32_t svqadd_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s64_z))) +svint64_t svqadd_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s16_z))) +svint16_t svqadd_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u8_m))) +svuint8_t svqadd_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u32_m))) +svuint32_t svqadd_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u64_m))) +svuint64_t svqadd_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u16_m))) +svuint16_t svqadd_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u8_x))) +svuint8_t svqadd_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u32_x))) +svuint32_t svqadd_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u64_x))) +svuint64_t svqadd_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u16_x))) +svuint16_t svqadd_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u8_z))) +svuint8_t svqadd_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u32_z))) +svuint32_t svqadd_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u64_z))) +svuint64_t svqadd_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u16_z))) +svuint16_t svqadd_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s8_m))) +svint8_t svqadd_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s32_m))) +svint32_t svqadd_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s64_m))) +svint64_t svqadd_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s16_m))) +svint16_t svqadd_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s8_x))) +svint8_t svqadd_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s32_x))) +svint32_t svqadd_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s64_x))) +svint64_t svqadd_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s16_x))) +svint16_t svqadd_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s8_z))) +svint8_t svqadd_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s32_z))) +svint32_t svqadd_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s64_z))) +svint64_t svqadd_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s16_z))) +svint16_t svqadd_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u8_m))) +svuint8_t svqadd_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u32_m))) +svuint32_t svqadd_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u64_m))) +svuint64_t svqadd_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u16_m))) +svuint16_t svqadd_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u8_x))) +svuint8_t svqadd_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u32_x))) +svuint32_t svqadd_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u64_x))) +svuint64_t svqadd_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u16_x))) +svuint16_t svqadd_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u8_z))) +svuint8_t svqadd_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u32_z))) +svuint32_t svqadd_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u64_z))) +svuint64_t svqadd_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u16_z))) +svuint16_t svqadd_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqcadd_s8))) +svint8_t svqcadd_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqcadd_s32))) +svint32_t svqcadd_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqcadd_s64))) +svint64_t svqcadd_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqcadd_s16))) +svint16_t svqcadd_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_n_s32))) +svint32_t svqdmlalb_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_n_s64))) +svint64_t svqdmlalb_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_n_s16))) +svint16_t svqdmlalb_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_s32))) +svint32_t svqdmlalb_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_s64))) +svint64_t svqdmlalb_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_s16))) +svint16_t svqdmlalb_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_lane_s32))) +svint32_t svqdmlalb_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_lane_s64))) +svint64_t svqdmlalb_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalbt_n_s32))) +svint32_t svqdmlalbt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalbt_n_s64))) +svint64_t svqdmlalbt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalbt_n_s16))) +svint16_t svqdmlalbt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalbt_s32))) +svint32_t svqdmlalbt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalbt_s64))) +svint64_t svqdmlalbt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalbt_s16))) +svint16_t svqdmlalbt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_n_s32))) +svint32_t svqdmlalt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_n_s64))) +svint64_t svqdmlalt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_n_s16))) +svint16_t svqdmlalt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_s32))) +svint32_t svqdmlalt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_s64))) +svint64_t svqdmlalt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_s16))) +svint16_t svqdmlalt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_lane_s32))) +svint32_t svqdmlalt_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_lane_s64))) +svint64_t svqdmlalt_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_n_s32))) +svint32_t svqdmlslb_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_n_s64))) +svint64_t svqdmlslb_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_n_s16))) +svint16_t svqdmlslb_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_s32))) +svint32_t svqdmlslb_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_s64))) +svint64_t svqdmlslb_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_s16))) +svint16_t svqdmlslb_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_lane_s32))) +svint32_t svqdmlslb_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_lane_s64))) +svint64_t svqdmlslb_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslbt_n_s32))) +svint32_t svqdmlslbt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslbt_n_s64))) +svint64_t svqdmlslbt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslbt_n_s16))) +svint16_t svqdmlslbt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslbt_s32))) +svint32_t svqdmlslbt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslbt_s64))) +svint64_t svqdmlslbt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslbt_s16))) +svint16_t svqdmlslbt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_n_s32))) +svint32_t svqdmlslt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_n_s64))) +svint64_t svqdmlslt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_n_s16))) +svint16_t svqdmlslt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_s32))) +svint32_t svqdmlslt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_s64))) +svint64_t svqdmlslt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_s16))) +svint16_t svqdmlslt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_lane_s32))) +svint32_t svqdmlslt_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_lane_s64))) +svint64_t svqdmlslt_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_n_s8))) +svint8_t svqdmulh_n_s8(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_n_s32))) +svint32_t svqdmulh_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_n_s64))) +svint64_t svqdmulh_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_n_s16))) +svint16_t svqdmulh_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_s8))) +svint8_t svqdmulh_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_s32))) +svint32_t svqdmulh_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_s64))) +svint64_t svqdmulh_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_s16))) +svint16_t svqdmulh_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_lane_s32))) +svint32_t svqdmulh_lane_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_lane_s64))) +svint64_t svqdmulh_lane_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_lane_s16))) +svint16_t svqdmulh_lane_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_n_s32))) +svint32_t svqdmullb_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_n_s64))) +svint64_t svqdmullb_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_n_s16))) +svint16_t svqdmullb_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_s32))) +svint32_t svqdmullb_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_s64))) +svint64_t svqdmullb_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_s16))) +svint16_t svqdmullb_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_lane_s32))) +svint32_t svqdmullb_lane_s32(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_lane_s64))) +svint64_t svqdmullb_lane_s64(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_n_s32))) +svint32_t svqdmullt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_n_s64))) +svint64_t svqdmullt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_n_s16))) +svint16_t svqdmullt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_s32))) +svint32_t svqdmullt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_s64))) +svint64_t svqdmullt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_s16))) +svint16_t svqdmullt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_lane_s32))) +svint32_t svqdmullt_lane_s32(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_lane_s64))) +svint64_t svqdmullt_lane_s64(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s8_m))) +svint8_t svqneg_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s32_m))) +svint32_t svqneg_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s64_m))) +svint64_t svqneg_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s16_m))) +svint16_t svqneg_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s8_x))) +svint8_t svqneg_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s32_x))) +svint32_t svqneg_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s64_x))) +svint64_t svqneg_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s16_x))) +svint16_t svqneg_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s8_z))) +svint8_t svqneg_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s32_z))) +svint32_t svqneg_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s64_z))) +svint64_t svqneg_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s16_z))) +svint16_t svqneg_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdcmlah_s8))) +svint8_t svqrdcmlah_s8(svint8_t, svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdcmlah_s32))) +svint32_t svqrdcmlah_s32(svint32_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdcmlah_s64))) +svint64_t svqrdcmlah_s64(svint64_t, svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdcmlah_s16))) +svint16_t svqrdcmlah_s16(svint16_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdcmlah_lane_s32))) +svint32_t svqrdcmlah_lane_s32(svint32_t, svint32_t, svint32_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdcmlah_lane_s16))) +svint16_t svqrdcmlah_lane_s16(svint16_t, svint16_t, svint16_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_n_s8))) +svint8_t svqrdmlah_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_n_s32))) +svint32_t svqrdmlah_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_n_s64))) +svint64_t svqrdmlah_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_n_s16))) +svint16_t svqrdmlah_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_s8))) +svint8_t svqrdmlah_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_s32))) +svint32_t svqrdmlah_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_s64))) +svint64_t svqrdmlah_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_s16))) +svint16_t svqrdmlah_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_lane_s32))) +svint32_t svqrdmlah_lane_s32(svint32_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_lane_s64))) +svint64_t svqrdmlah_lane_s64(svint64_t, svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_lane_s16))) +svint16_t svqrdmlah_lane_s16(svint16_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_n_s8))) +svint8_t svqrdmlsh_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_n_s32))) +svint32_t svqrdmlsh_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_n_s64))) +svint64_t svqrdmlsh_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_n_s16))) +svint16_t svqrdmlsh_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_s8))) +svint8_t svqrdmlsh_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_s32))) +svint32_t svqrdmlsh_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_s64))) +svint64_t svqrdmlsh_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_s16))) +svint16_t svqrdmlsh_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_lane_s32))) +svint32_t svqrdmlsh_lane_s32(svint32_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_lane_s64))) +svint64_t svqrdmlsh_lane_s64(svint64_t, svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_lane_s16))) +svint16_t svqrdmlsh_lane_s16(svint16_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_n_s8))) +svint8_t svqrdmulh_n_s8(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_n_s32))) +svint32_t svqrdmulh_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_n_s64))) +svint64_t svqrdmulh_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_n_s16))) +svint16_t svqrdmulh_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_s8))) +svint8_t svqrdmulh_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_s32))) +svint32_t svqrdmulh_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_s64))) +svint64_t svqrdmulh_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_s16))) +svint16_t svqrdmulh_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_lane_s32))) +svint32_t svqrdmulh_lane_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_lane_s64))) +svint64_t svqrdmulh_lane_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_lane_s16))) +svint16_t svqrdmulh_lane_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s8_m))) +svint8_t svqrshl_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s32_m))) +svint32_t svqrshl_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s64_m))) +svint64_t svqrshl_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s16_m))) +svint16_t svqrshl_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s8_x))) +svint8_t svqrshl_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s32_x))) +svint32_t svqrshl_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s64_x))) +svint64_t svqrshl_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s16_x))) +svint16_t svqrshl_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s8_z))) +svint8_t svqrshl_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s32_z))) +svint32_t svqrshl_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s64_z))) +svint64_t svqrshl_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s16_z))) +svint16_t svqrshl_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u8_m))) +svuint8_t svqrshl_n_u8_m(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u32_m))) +svuint32_t svqrshl_n_u32_m(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u64_m))) +svuint64_t svqrshl_n_u64_m(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u16_m))) +svuint16_t svqrshl_n_u16_m(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u8_x))) +svuint8_t svqrshl_n_u8_x(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u32_x))) +svuint32_t svqrshl_n_u32_x(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u64_x))) +svuint64_t svqrshl_n_u64_x(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u16_x))) +svuint16_t svqrshl_n_u16_x(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u8_z))) +svuint8_t svqrshl_n_u8_z(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u32_z))) +svuint32_t svqrshl_n_u32_z(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u64_z))) +svuint64_t svqrshl_n_u64_z(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u16_z))) +svuint16_t svqrshl_n_u16_z(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s8_m))) +svint8_t svqrshl_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s32_m))) +svint32_t svqrshl_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s64_m))) +svint64_t svqrshl_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s16_m))) +svint16_t svqrshl_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s8_x))) +svint8_t svqrshl_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s32_x))) +svint32_t svqrshl_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s64_x))) +svint64_t svqrshl_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s16_x))) +svint16_t svqrshl_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s8_z))) +svint8_t svqrshl_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s32_z))) +svint32_t svqrshl_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s64_z))) +svint64_t svqrshl_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s16_z))) +svint16_t svqrshl_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u8_m))) +svuint8_t svqrshl_u8_m(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u32_m))) +svuint32_t svqrshl_u32_m(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u64_m))) +svuint64_t svqrshl_u64_m(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u16_m))) +svuint16_t svqrshl_u16_m(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u8_x))) +svuint8_t svqrshl_u8_x(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u32_x))) +svuint32_t svqrshl_u32_x(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u64_x))) +svuint64_t svqrshl_u64_x(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u16_x))) +svuint16_t svqrshl_u16_x(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u8_z))) +svuint8_t svqrshl_u8_z(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u32_z))) +svuint32_t svqrshl_u32_z(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u64_z))) +svuint64_t svqrshl_u64_z(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u16_z))) +svuint16_t svqrshl_u16_z(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnb_n_s32))) +svint16_t svqrshrnb_n_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnb_n_s64))) +svint32_t svqrshrnb_n_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnb_n_s16))) +svint8_t svqrshrnb_n_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnb_n_u32))) +svuint16_t svqrshrnb_n_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnb_n_u64))) +svuint32_t svqrshrnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnb_n_u16))) +svuint8_t svqrshrnb_n_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnt_n_s32))) +svint16_t svqrshrnt_n_s32(svint16_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnt_n_s64))) +svint32_t svqrshrnt_n_s64(svint32_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnt_n_s16))) +svint8_t svqrshrnt_n_s16(svint8_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnt_n_u32))) +svuint16_t svqrshrnt_n_u32(svuint16_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnt_n_u64))) +svuint32_t svqrshrnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnt_n_u16))) +svuint8_t svqrshrnt_n_u16(svuint8_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrunb_n_s32))) +svuint16_t svqrshrunb_n_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrunb_n_s64))) +svuint32_t svqrshrunb_n_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrunb_n_s16))) +svuint8_t svqrshrunb_n_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrunt_n_s32))) +svuint16_t svqrshrunt_n_s32(svuint16_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrunt_n_s64))) +svuint32_t svqrshrunt_n_s64(svuint32_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrunt_n_s16))) +svuint8_t svqrshrunt_n_s16(svuint8_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s8_m))) +svint8_t svqshl_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s32_m))) +svint32_t svqshl_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s64_m))) +svint64_t svqshl_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s16_m))) +svint16_t svqshl_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s8_x))) +svint8_t svqshl_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s32_x))) +svint32_t svqshl_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s64_x))) +svint64_t svqshl_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s16_x))) +svint16_t svqshl_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s8_z))) +svint8_t svqshl_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s32_z))) +svint32_t svqshl_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s64_z))) +svint64_t svqshl_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s16_z))) +svint16_t svqshl_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u8_m))) +svuint8_t svqshl_n_u8_m(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u32_m))) +svuint32_t svqshl_n_u32_m(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u64_m))) +svuint64_t svqshl_n_u64_m(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u16_m))) +svuint16_t svqshl_n_u16_m(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u8_x))) +svuint8_t svqshl_n_u8_x(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u32_x))) +svuint32_t svqshl_n_u32_x(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u64_x))) +svuint64_t svqshl_n_u64_x(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u16_x))) +svuint16_t svqshl_n_u16_x(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u8_z))) +svuint8_t svqshl_n_u8_z(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u32_z))) +svuint32_t svqshl_n_u32_z(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u64_z))) +svuint64_t svqshl_n_u64_z(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u16_z))) +svuint16_t svqshl_n_u16_z(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s8_m))) +svint8_t svqshl_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s32_m))) +svint32_t svqshl_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s64_m))) +svint64_t svqshl_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s16_m))) +svint16_t svqshl_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s8_x))) +svint8_t svqshl_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s32_x))) +svint32_t svqshl_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s64_x))) +svint64_t svqshl_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s16_x))) +svint16_t svqshl_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s8_z))) +svint8_t svqshl_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s32_z))) +svint32_t svqshl_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s64_z))) +svint64_t svqshl_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s16_z))) +svint16_t svqshl_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u8_m))) +svuint8_t svqshl_u8_m(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u32_m))) +svuint32_t svqshl_u32_m(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u64_m))) +svuint64_t svqshl_u64_m(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u16_m))) +svuint16_t svqshl_u16_m(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u8_x))) +svuint8_t svqshl_u8_x(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u32_x))) +svuint32_t svqshl_u32_x(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u64_x))) +svuint64_t svqshl_u64_x(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u16_x))) +svuint16_t svqshl_u16_x(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u8_z))) +svuint8_t svqshl_u8_z(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u32_z))) +svuint32_t svqshl_u32_z(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u64_z))) +svuint64_t svqshl_u64_z(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u16_z))) +svuint16_t svqshl_u16_z(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s8_m))) +svuint8_t svqshlu_n_s8_m(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s32_m))) +svuint32_t svqshlu_n_s32_m(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s64_m))) +svuint64_t svqshlu_n_s64_m(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s16_m))) +svuint16_t svqshlu_n_s16_m(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s8_x))) +svuint8_t svqshlu_n_s8_x(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s32_x))) +svuint32_t svqshlu_n_s32_x(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s64_x))) +svuint64_t svqshlu_n_s64_x(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s16_x))) +svuint16_t svqshlu_n_s16_x(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s8_z))) +svuint8_t svqshlu_n_s8_z(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s32_z))) +svuint32_t svqshlu_n_s32_z(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s64_z))) +svuint64_t svqshlu_n_s64_z(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s16_z))) +svuint16_t svqshlu_n_s16_z(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnb_n_s32))) +svint16_t svqshrnb_n_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnb_n_s64))) +svint32_t svqshrnb_n_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnb_n_s16))) +svint8_t svqshrnb_n_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnb_n_u32))) +svuint16_t svqshrnb_n_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnb_n_u64))) +svuint32_t svqshrnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnb_n_u16))) +svuint8_t svqshrnb_n_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnt_n_s32))) +svint16_t svqshrnt_n_s32(svint16_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnt_n_s64))) +svint32_t svqshrnt_n_s64(svint32_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnt_n_s16))) +svint8_t svqshrnt_n_s16(svint8_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnt_n_u32))) +svuint16_t svqshrnt_n_u32(svuint16_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnt_n_u64))) +svuint32_t svqshrnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnt_n_u16))) +svuint8_t svqshrnt_n_u16(svuint8_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrunb_n_s32))) +svuint16_t svqshrunb_n_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrunb_n_s64))) +svuint32_t svqshrunb_n_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrunb_n_s16))) +svuint8_t svqshrunb_n_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrunt_n_s32))) +svuint16_t svqshrunt_n_s32(svuint16_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrunt_n_s64))) +svuint32_t svqshrunt_n_s64(svuint32_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrunt_n_s16))) +svuint8_t svqshrunt_n_s16(svuint8_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s8_m))) +svint8_t svqsub_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s32_m))) +svint32_t svqsub_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s64_m))) +svint64_t svqsub_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s16_m))) +svint16_t svqsub_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s8_x))) +svint8_t svqsub_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s32_x))) +svint32_t svqsub_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s64_x))) +svint64_t svqsub_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s16_x))) +svint16_t svqsub_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s8_z))) +svint8_t svqsub_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s32_z))) +svint32_t svqsub_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s64_z))) +svint64_t svqsub_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s16_z))) +svint16_t svqsub_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u8_m))) +svuint8_t svqsub_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u32_m))) +svuint32_t svqsub_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u64_m))) +svuint64_t svqsub_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u16_m))) +svuint16_t svqsub_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u8_x))) +svuint8_t svqsub_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u32_x))) +svuint32_t svqsub_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u64_x))) +svuint64_t svqsub_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u16_x))) +svuint16_t svqsub_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u8_z))) +svuint8_t svqsub_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u32_z))) +svuint32_t svqsub_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u64_z))) +svuint64_t svqsub_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u16_z))) +svuint16_t svqsub_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s8_m))) +svint8_t svqsub_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s32_m))) +svint32_t svqsub_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s64_m))) +svint64_t svqsub_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s16_m))) +svint16_t svqsub_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s8_x))) +svint8_t svqsub_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s32_x))) +svint32_t svqsub_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s64_x))) +svint64_t svqsub_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s16_x))) +svint16_t svqsub_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s8_z))) +svint8_t svqsub_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s32_z))) +svint32_t svqsub_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s64_z))) +svint64_t svqsub_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s16_z))) +svint16_t svqsub_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u8_m))) +svuint8_t svqsub_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u32_m))) +svuint32_t svqsub_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u64_m))) +svuint64_t svqsub_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u16_m))) +svuint16_t svqsub_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u8_x))) +svuint8_t svqsub_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u32_x))) +svuint32_t svqsub_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u64_x))) +svuint64_t svqsub_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u16_x))) +svuint16_t svqsub_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u8_z))) +svuint8_t svqsub_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u32_z))) +svuint32_t svqsub_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u64_z))) +svuint64_t svqsub_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u16_z))) +svuint16_t svqsub_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s8_m))) +svint8_t svqsubr_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s32_m))) +svint32_t svqsubr_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s64_m))) +svint64_t svqsubr_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s16_m))) +svint16_t svqsubr_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s8_x))) +svint8_t svqsubr_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s32_x))) +svint32_t svqsubr_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s64_x))) +svint64_t svqsubr_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s16_x))) +svint16_t svqsubr_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s8_z))) +svint8_t svqsubr_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s32_z))) +svint32_t svqsubr_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s64_z))) +svint64_t svqsubr_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s16_z))) +svint16_t svqsubr_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u8_m))) +svuint8_t svqsubr_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u32_m))) +svuint32_t svqsubr_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u64_m))) +svuint64_t svqsubr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u16_m))) +svuint16_t svqsubr_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u8_x))) +svuint8_t svqsubr_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u32_x))) +svuint32_t svqsubr_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u64_x))) +svuint64_t svqsubr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u16_x))) +svuint16_t svqsubr_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u8_z))) +svuint8_t svqsubr_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u32_z))) +svuint32_t svqsubr_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u64_z))) +svuint64_t svqsubr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u16_z))) +svuint16_t svqsubr_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s8_m))) +svint8_t svqsubr_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s32_m))) +svint32_t svqsubr_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s64_m))) +svint64_t svqsubr_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s16_m))) +svint16_t svqsubr_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s8_x))) +svint8_t svqsubr_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s32_x))) +svint32_t svqsubr_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s64_x))) +svint64_t svqsubr_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s16_x))) +svint16_t svqsubr_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s8_z))) +svint8_t svqsubr_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s32_z))) +svint32_t svqsubr_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s64_z))) +svint64_t svqsubr_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s16_z))) +svint16_t svqsubr_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u8_m))) +svuint8_t svqsubr_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u32_m))) +svuint32_t svqsubr_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u64_m))) +svuint64_t svqsubr_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u16_m))) +svuint16_t svqsubr_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u8_x))) +svuint8_t svqsubr_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u32_x))) +svuint32_t svqsubr_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u64_x))) +svuint64_t svqsubr_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u16_x))) +svuint16_t svqsubr_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u8_z))) +svuint8_t svqsubr_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u32_z))) +svuint32_t svqsubr_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u64_z))) +svuint64_t svqsubr_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u16_z))) +svuint16_t svqsubr_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnb_s32))) +svint16_t svqxtnb_s32(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnb_s64))) +svint32_t svqxtnb_s64(svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnb_s16))) +svint8_t svqxtnb_s16(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnb_u32))) +svuint16_t svqxtnb_u32(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnb_u64))) +svuint32_t svqxtnb_u64(svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnb_u16))) +svuint8_t svqxtnb_u16(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnt_s32))) +svint16_t svqxtnt_s32(svint16_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnt_s64))) +svint32_t svqxtnt_s64(svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnt_s16))) +svint8_t svqxtnt_s16(svint8_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnt_u32))) +svuint16_t svqxtnt_u32(svuint16_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnt_u64))) +svuint32_t svqxtnt_u64(svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnt_u16))) +svuint8_t svqxtnt_u16(svuint8_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtunb_s32))) +svuint16_t svqxtunb_s32(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtunb_s64))) +svuint32_t svqxtunb_s64(svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtunb_s16))) +svuint8_t svqxtunb_s16(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtunt_s32))) +svuint16_t svqxtunt_s32(svuint16_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtunt_s64))) +svuint32_t svqxtunt_s64(svuint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtunt_s16))) +svuint8_t svqxtunt_s16(svuint8_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_n_u32))) +svuint16_t svraddhnb_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_n_u64))) +svuint32_t svraddhnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_n_u16))) +svuint8_t svraddhnb_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_n_s32))) +svint16_t svraddhnb_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_n_s64))) +svint32_t svraddhnb_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_n_s16))) +svint8_t svraddhnb_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_u32))) +svuint16_t svraddhnb_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_u64))) +svuint32_t svraddhnb_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_u16))) +svuint8_t svraddhnb_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_s32))) +svint16_t svraddhnb_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_s64))) +svint32_t svraddhnb_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_s16))) +svint8_t svraddhnb_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_n_u32))) +svuint16_t svraddhnt_n_u32(svuint16_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_n_u64))) +svuint32_t svraddhnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_n_u16))) +svuint8_t svraddhnt_n_u16(svuint8_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_n_s32))) +svint16_t svraddhnt_n_s32(svint16_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_n_s64))) +svint32_t svraddhnt_n_s64(svint32_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_n_s16))) +svint8_t svraddhnt_n_s16(svint8_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_u32))) +svuint16_t svraddhnt_u32(svuint16_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_u64))) +svuint32_t svraddhnt_u64(svuint32_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_u16))) +svuint8_t svraddhnt_u16(svuint8_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_s32))) +svint16_t svraddhnt_s32(svint16_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_s64))) +svint32_t svraddhnt_s64(svint32_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_s16))) +svint8_t svraddhnt_s16(svint8_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpe_u32_m))) +svuint32_t svrecpe_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpe_u32_x))) +svuint32_t svrecpe_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpe_u32_z))) +svuint32_t svrecpe_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s8_m))) +svint8_t svrhadd_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s32_m))) +svint32_t svrhadd_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s64_m))) +svint64_t svrhadd_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s16_m))) +svint16_t svrhadd_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s8_x))) +svint8_t svrhadd_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s32_x))) +svint32_t svrhadd_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s64_x))) +svint64_t svrhadd_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s16_x))) +svint16_t svrhadd_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s8_z))) +svint8_t svrhadd_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s32_z))) +svint32_t svrhadd_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s64_z))) +svint64_t svrhadd_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s16_z))) +svint16_t svrhadd_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u8_m))) +svuint8_t svrhadd_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u32_m))) +svuint32_t svrhadd_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u64_m))) +svuint64_t svrhadd_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u16_m))) +svuint16_t svrhadd_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u8_x))) +svuint8_t svrhadd_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u32_x))) +svuint32_t svrhadd_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u64_x))) +svuint64_t svrhadd_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u16_x))) +svuint16_t svrhadd_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u8_z))) +svuint8_t svrhadd_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u32_z))) +svuint32_t svrhadd_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u64_z))) +svuint64_t svrhadd_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u16_z))) +svuint16_t svrhadd_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s8_m))) +svint8_t svrhadd_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s32_m))) +svint32_t svrhadd_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s64_m))) +svint64_t svrhadd_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s16_m))) +svint16_t svrhadd_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s8_x))) +svint8_t svrhadd_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s32_x))) +svint32_t svrhadd_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s64_x))) +svint64_t svrhadd_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s16_x))) +svint16_t svrhadd_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s8_z))) +svint8_t svrhadd_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s32_z))) +svint32_t svrhadd_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s64_z))) +svint64_t svrhadd_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s16_z))) +svint16_t svrhadd_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u8_m))) +svuint8_t svrhadd_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u32_m))) +svuint32_t svrhadd_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u64_m))) +svuint64_t svrhadd_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u16_m))) +svuint16_t svrhadd_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u8_x))) +svuint8_t svrhadd_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u32_x))) +svuint32_t svrhadd_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u64_x))) +svuint64_t svrhadd_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u16_x))) +svuint16_t svrhadd_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u8_z))) +svuint8_t svrhadd_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u32_z))) +svuint32_t svrhadd_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u64_z))) +svuint64_t svrhadd_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u16_z))) +svuint16_t svrhadd_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s8_m))) +svint8_t svrshl_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s32_m))) +svint32_t svrshl_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s64_m))) +svint64_t svrshl_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s16_m))) +svint16_t svrshl_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s8_x))) +svint8_t svrshl_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s32_x))) +svint32_t svrshl_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s64_x))) +svint64_t svrshl_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s16_x))) +svint16_t svrshl_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s8_z))) +svint8_t svrshl_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s32_z))) +svint32_t svrshl_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s64_z))) +svint64_t svrshl_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s16_z))) +svint16_t svrshl_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u8_m))) +svuint8_t svrshl_n_u8_m(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u32_m))) +svuint32_t svrshl_n_u32_m(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u64_m))) +svuint64_t svrshl_n_u64_m(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u16_m))) +svuint16_t svrshl_n_u16_m(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u8_x))) +svuint8_t svrshl_n_u8_x(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u32_x))) +svuint32_t svrshl_n_u32_x(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u64_x))) +svuint64_t svrshl_n_u64_x(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u16_x))) +svuint16_t svrshl_n_u16_x(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u8_z))) +svuint8_t svrshl_n_u8_z(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u32_z))) +svuint32_t svrshl_n_u32_z(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u64_z))) +svuint64_t svrshl_n_u64_z(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u16_z))) +svuint16_t svrshl_n_u16_z(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s8_m))) +svint8_t svrshl_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s32_m))) +svint32_t svrshl_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s64_m))) +svint64_t svrshl_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s16_m))) +svint16_t svrshl_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s8_x))) +svint8_t svrshl_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s32_x))) +svint32_t svrshl_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s64_x))) +svint64_t svrshl_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s16_x))) +svint16_t svrshl_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s8_z))) +svint8_t svrshl_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s32_z))) +svint32_t svrshl_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s64_z))) +svint64_t svrshl_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s16_z))) +svint16_t svrshl_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u8_m))) +svuint8_t svrshl_u8_m(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u32_m))) +svuint32_t svrshl_u32_m(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u64_m))) +svuint64_t svrshl_u64_m(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u16_m))) +svuint16_t svrshl_u16_m(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u8_x))) +svuint8_t svrshl_u8_x(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u32_x))) +svuint32_t svrshl_u32_x(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u64_x))) +svuint64_t svrshl_u64_x(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u16_x))) +svuint16_t svrshl_u16_x(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u8_z))) +svuint8_t svrshl_u8_z(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u32_z))) +svuint32_t svrshl_u32_z(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u64_z))) +svuint64_t svrshl_u64_z(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u16_z))) +svuint16_t svrshl_u16_z(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s8_m))) +svint8_t svrshr_n_s8_m(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s32_m))) +svint32_t svrshr_n_s32_m(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s64_m))) +svint64_t svrshr_n_s64_m(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s16_m))) +svint16_t svrshr_n_s16_m(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u8_m))) +svuint8_t svrshr_n_u8_m(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u32_m))) +svuint32_t svrshr_n_u32_m(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u64_m))) +svuint64_t svrshr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u16_m))) +svuint16_t svrshr_n_u16_m(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s8_x))) +svint8_t svrshr_n_s8_x(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s32_x))) +svint32_t svrshr_n_s32_x(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s64_x))) +svint64_t svrshr_n_s64_x(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s16_x))) +svint16_t svrshr_n_s16_x(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u8_x))) +svuint8_t svrshr_n_u8_x(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u32_x))) +svuint32_t svrshr_n_u32_x(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u64_x))) +svuint64_t svrshr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u16_x))) +svuint16_t svrshr_n_u16_x(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s8_z))) +svint8_t svrshr_n_s8_z(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s32_z))) +svint32_t svrshr_n_s32_z(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s64_z))) +svint64_t svrshr_n_s64_z(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s16_z))) +svint16_t svrshr_n_s16_z(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u8_z))) +svuint8_t svrshr_n_u8_z(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u32_z))) +svuint32_t svrshr_n_u32_z(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u64_z))) +svuint64_t svrshr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u16_z))) +svuint16_t svrshr_n_u16_z(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnb_n_u32))) +svuint16_t svrshrnb_n_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnb_n_u64))) +svuint32_t svrshrnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnb_n_u16))) +svuint8_t svrshrnb_n_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnb_n_s32))) +svint16_t svrshrnb_n_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnb_n_s64))) +svint32_t svrshrnb_n_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnb_n_s16))) +svint8_t svrshrnb_n_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnt_n_u32))) +svuint16_t svrshrnt_n_u32(svuint16_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnt_n_u64))) +svuint32_t svrshrnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnt_n_u16))) +svuint8_t svrshrnt_n_u16(svuint8_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnt_n_s32))) +svint16_t svrshrnt_n_s32(svint16_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnt_n_s64))) +svint32_t svrshrnt_n_s64(svint32_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnt_n_s16))) +svint8_t svrshrnt_n_s16(svint8_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrte_u32_m))) +svuint32_t svrsqrte_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrte_u32_x))) +svuint32_t svrsqrte_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrte_u32_z))) +svuint32_t svrsqrte_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_s8))) +svint8_t svrsra_n_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_s32))) +svint32_t svrsra_n_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_s64))) +svint64_t svrsra_n_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_s16))) +svint16_t svrsra_n_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_u8))) +svuint8_t svrsra_n_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_u32))) +svuint32_t svrsra_n_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_u64))) +svuint64_t svrsra_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_u16))) +svuint16_t svrsra_n_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_n_u32))) +svuint16_t svrsubhnb_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_n_u64))) +svuint32_t svrsubhnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_n_u16))) +svuint8_t svrsubhnb_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_n_s32))) +svint16_t svrsubhnb_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_n_s64))) +svint32_t svrsubhnb_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_n_s16))) +svint8_t svrsubhnb_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_u32))) +svuint16_t svrsubhnb_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_u64))) +svuint32_t svrsubhnb_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_u16))) +svuint8_t svrsubhnb_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_s32))) +svint16_t svrsubhnb_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_s64))) +svint32_t svrsubhnb_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_s16))) +svint8_t svrsubhnb_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_n_u32))) +svuint16_t svrsubhnt_n_u32(svuint16_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_n_u64))) +svuint32_t svrsubhnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_n_u16))) +svuint8_t svrsubhnt_n_u16(svuint8_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_n_s32))) +svint16_t svrsubhnt_n_s32(svint16_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_n_s64))) +svint32_t svrsubhnt_n_s64(svint32_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_n_s16))) +svint8_t svrsubhnt_n_s16(svint8_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_u32))) +svuint16_t svrsubhnt_u32(svuint16_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_u64))) +svuint32_t svrsubhnt_u64(svuint32_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_u16))) +svuint8_t svrsubhnt_u16(svuint8_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_s32))) +svint16_t svrsubhnt_s32(svint16_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_s64))) +svint32_t svrsubhnt_s64(svint32_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_s16))) +svint8_t svrsubhnt_s16(svint8_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclb_n_u32))) +svuint32_t svsbclb_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclb_n_u64))) +svuint64_t svsbclb_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclb_u32))) +svuint32_t svsbclb_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclb_u64))) +svuint64_t svsbclb_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclt_n_u32))) +svuint32_t svsbclt_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclt_n_u64))) +svuint64_t svsbclt_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclt_u32))) +svuint32_t svsbclt_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclt_u64))) +svuint64_t svsbclt_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllb_n_s32))) +svint32_t svshllb_n_s32(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllb_n_s64))) +svint64_t svshllb_n_s64(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllb_n_s16))) +svint16_t svshllb_n_s16(svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllb_n_u32))) +svuint32_t svshllb_n_u32(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllb_n_u64))) +svuint64_t svshllb_n_u64(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllb_n_u16))) +svuint16_t svshllb_n_u16(svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllt_n_s32))) +svint32_t svshllt_n_s32(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllt_n_s64))) +svint64_t svshllt_n_s64(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllt_n_s16))) +svint16_t svshllt_n_s16(svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllt_n_u32))) +svuint32_t svshllt_n_u32(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllt_n_u64))) +svuint64_t svshllt_n_u64(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllt_n_u16))) +svuint16_t svshllt_n_u16(svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnb_n_u32))) +svuint16_t svshrnb_n_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnb_n_u64))) +svuint32_t svshrnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnb_n_u16))) +svuint8_t svshrnb_n_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnb_n_s32))) +svint16_t svshrnb_n_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnb_n_s64))) +svint32_t svshrnb_n_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnb_n_s16))) +svint8_t svshrnb_n_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnt_n_u32))) +svuint16_t svshrnt_n_u32(svuint16_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnt_n_u64))) +svuint32_t svshrnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnt_n_u16))) +svuint8_t svshrnt_n_u16(svuint8_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnt_n_s32))) +svint16_t svshrnt_n_s32(svint16_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnt_n_s64))) +svint32_t svshrnt_n_s64(svint32_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnt_n_s16))) +svint8_t svshrnt_n_s16(svint8_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_u8))) +svuint8_t svsli_n_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_u32))) +svuint32_t svsli_n_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_u64))) +svuint64_t svsli_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_u16))) +svuint16_t svsli_n_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_s8))) +svint8_t svsli_n_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_s32))) +svint32_t svsli_n_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_s64))) +svint64_t svsli_n_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_s16))) +svint16_t svsli_n_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u8_m))) +svuint8_t svsqadd_n_u8_m(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u32_m))) +svuint32_t svsqadd_n_u32_m(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u64_m))) +svuint64_t svsqadd_n_u64_m(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u16_m))) +svuint16_t svsqadd_n_u16_m(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u8_x))) +svuint8_t svsqadd_n_u8_x(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u32_x))) +svuint32_t svsqadd_n_u32_x(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u64_x))) +svuint64_t svsqadd_n_u64_x(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u16_x))) +svuint16_t svsqadd_n_u16_x(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u8_z))) +svuint8_t svsqadd_n_u8_z(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u32_z))) +svuint32_t svsqadd_n_u32_z(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u64_z))) +svuint64_t svsqadd_n_u64_z(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u16_z))) +svuint16_t svsqadd_n_u16_z(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u8_m))) +svuint8_t svsqadd_u8_m(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u32_m))) +svuint32_t svsqadd_u32_m(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u64_m))) +svuint64_t svsqadd_u64_m(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u16_m))) +svuint16_t svsqadd_u16_m(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u8_x))) +svuint8_t svsqadd_u8_x(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u32_x))) +svuint32_t svsqadd_u32_x(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u64_x))) +svuint64_t svsqadd_u64_x(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u16_x))) +svuint16_t svsqadd_u16_x(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u8_z))) +svuint8_t svsqadd_u8_z(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u32_z))) +svuint32_t svsqadd_u32_z(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u64_z))) +svuint64_t svsqadd_u64_z(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u16_z))) +svuint16_t svsqadd_u16_z(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_s8))) +svint8_t svsra_n_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_s32))) +svint32_t svsra_n_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_s64))) +svint64_t svsra_n_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_s16))) +svint16_t svsra_n_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_u8))) +svuint8_t svsra_n_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_u32))) +svuint32_t svsra_n_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_u64))) +svuint64_t svsra_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_u16))) +svuint16_t svsra_n_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_u8))) +svuint8_t svsri_n_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_u32))) +svuint32_t svsri_n_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_u64))) +svuint64_t svsri_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_u16))) +svuint16_t svsri_n_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_s8))) +svint8_t svsri_n_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_s32))) +svint32_t svsri_n_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_s64))) +svint64_t svsri_n_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_s16))) +svint16_t svsri_n_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_index_u32))) +void svstnt1_scatter_u32base_index_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_index_u64))) +void svstnt1_scatter_u64base_index_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_index_f64))) +void svstnt1_scatter_u64base_index_f64(svbool_t, svuint64_t, int64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_index_f32))) +void svstnt1_scatter_u32base_index_f32(svbool_t, svuint32_t, int64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_index_s32))) +void svstnt1_scatter_u32base_index_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_index_s64))) +void svstnt1_scatter_u64base_index_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_offset_u32))) +void svstnt1_scatter_u32base_offset_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_offset_u64))) +void svstnt1_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_offset_f64))) +void svstnt1_scatter_u64base_offset_f64(svbool_t, svuint64_t, int64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_offset_f32))) +void svstnt1_scatter_u32base_offset_f32(svbool_t, svuint32_t, int64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_offset_s32))) +void svstnt1_scatter_u32base_offset_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_offset_s64))) +void svstnt1_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_u32))) +void svstnt1_scatter_u32base_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_u64))) +void svstnt1_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_f64))) +void svstnt1_scatter_u64base_f64(svbool_t, svuint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_f32))) +void svstnt1_scatter_u32base_f32(svbool_t, svuint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_s32))) +void svstnt1_scatter_u32base_s32(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_s64))) +void svstnt1_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_s64index_u64))) +void svstnt1_scatter_s64index_u64(svbool_t, uint64_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_s64index_f64))) +void svstnt1_scatter_s64index_f64(svbool_t, float64_t *, svint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_s64index_s64))) +void svstnt1_scatter_s64index_s64(svbool_t, int64_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64index_u64))) +void svstnt1_scatter_u64index_u64(svbool_t, uint64_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64index_f64))) +void svstnt1_scatter_u64index_f64(svbool_t, float64_t *, svuint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64index_s64))) +void svstnt1_scatter_u64index_s64(svbool_t, int64_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32offset_u32))) +void svstnt1_scatter_u32offset_u32(svbool_t, uint32_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32offset_f32))) +void svstnt1_scatter_u32offset_f32(svbool_t, float32_t *, svuint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32offset_s32))) +void svstnt1_scatter_u32offset_s32(svbool_t, int32_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_s64offset_u64))) +void svstnt1_scatter_s64offset_u64(svbool_t, uint64_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_s64offset_f64))) +void svstnt1_scatter_s64offset_f64(svbool_t, float64_t *, svint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_s64offset_s64))) +void svstnt1_scatter_s64offset_s64(svbool_t, int64_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64offset_u64))) +void svstnt1_scatter_u64offset_u64(svbool_t, uint64_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64offset_f64))) +void svstnt1_scatter_u64offset_f64(svbool_t, float64_t *, svuint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64offset_s64))) +void svstnt1_scatter_u64offset_s64(svbool_t, int64_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u32base_offset_u32))) +void svstnt1b_scatter_u32base_offset_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u64base_offset_u64))) +void svstnt1b_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u32base_offset_s32))) +void svstnt1b_scatter_u32base_offset_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u64base_offset_s64))) +void svstnt1b_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u32base_u32))) +void svstnt1b_scatter_u32base_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u64base_u64))) +void svstnt1b_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u32base_s32))) +void svstnt1b_scatter_u32base_s32(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u64base_s64))) +void svstnt1b_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u32offset_s32))) +void svstnt1b_scatter_u32offset_s32(svbool_t, int8_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u32offset_u32))) +void svstnt1b_scatter_u32offset_u32(svbool_t, uint8_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_s64offset_s64))) +void svstnt1b_scatter_s64offset_s64(svbool_t, int8_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_s64offset_u64))) +void svstnt1b_scatter_s64offset_u64(svbool_t, uint8_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u64offset_s64))) +void svstnt1b_scatter_u64offset_s64(svbool_t, int8_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u64offset_u64))) +void svstnt1b_scatter_u64offset_u64(svbool_t, uint8_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32base_index_u32))) +void svstnt1h_scatter_u32base_index_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64base_index_u64))) +void svstnt1h_scatter_u64base_index_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32base_index_s32))) +void svstnt1h_scatter_u32base_index_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64base_index_s64))) +void svstnt1h_scatter_u64base_index_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32base_offset_u32))) +void svstnt1h_scatter_u32base_offset_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64base_offset_u64))) +void svstnt1h_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32base_offset_s32))) +void svstnt1h_scatter_u32base_offset_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64base_offset_s64))) +void svstnt1h_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32base_u32))) +void svstnt1h_scatter_u32base_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64base_u64))) +void svstnt1h_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32base_s32))) +void svstnt1h_scatter_u32base_s32(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64base_s64))) +void svstnt1h_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_s64index_s64))) +void svstnt1h_scatter_s64index_s64(svbool_t, int16_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_s64index_u64))) +void svstnt1h_scatter_s64index_u64(svbool_t, uint16_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64index_s64))) +void svstnt1h_scatter_u64index_s64(svbool_t, int16_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64index_u64))) +void svstnt1h_scatter_u64index_u64(svbool_t, uint16_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32offset_s32))) +void svstnt1h_scatter_u32offset_s32(svbool_t, int16_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32offset_u32))) +void svstnt1h_scatter_u32offset_u32(svbool_t, uint16_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_s64offset_s64))) +void svstnt1h_scatter_s64offset_s64(svbool_t, int16_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_s64offset_u64))) +void svstnt1h_scatter_s64offset_u64(svbool_t, uint16_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64offset_s64))) +void svstnt1h_scatter_u64offset_s64(svbool_t, int16_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64offset_u64))) +void svstnt1h_scatter_u64offset_u64(svbool_t, uint16_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64base_index_u64))) +void svstnt1w_scatter_u64base_index_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64base_index_s64))) +void svstnt1w_scatter_u64base_index_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64base_offset_u64))) +void svstnt1w_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64base_offset_s64))) +void svstnt1w_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64base_u64))) +void svstnt1w_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64base_s64))) +void svstnt1w_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_s64index_s64))) +void svstnt1w_scatter_s64index_s64(svbool_t, int32_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_s64index_u64))) +void svstnt1w_scatter_s64index_u64(svbool_t, uint32_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64index_s64))) +void svstnt1w_scatter_u64index_s64(svbool_t, int32_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64index_u64))) +void svstnt1w_scatter_u64index_u64(svbool_t, uint32_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_s64offset_s64))) +void svstnt1w_scatter_s64offset_s64(svbool_t, int32_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_s64offset_u64))) +void svstnt1w_scatter_s64offset_u64(svbool_t, uint32_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64offset_s64))) +void svstnt1w_scatter_u64offset_s64(svbool_t, int32_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64offset_u64))) +void svstnt1w_scatter_u64offset_u64(svbool_t, uint32_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_n_u32))) +svuint16_t svsubhnb_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_n_u64))) +svuint32_t svsubhnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_n_u16))) +svuint8_t svsubhnb_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_n_s32))) +svint16_t svsubhnb_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_n_s64))) +svint32_t svsubhnb_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_n_s16))) +svint8_t svsubhnb_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_u32))) +svuint16_t svsubhnb_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_u64))) +svuint32_t svsubhnb_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_u16))) +svuint8_t svsubhnb_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_s32))) +svint16_t svsubhnb_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_s64))) +svint32_t svsubhnb_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_s16))) +svint8_t svsubhnb_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_n_u32))) +svuint16_t svsubhnt_n_u32(svuint16_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_n_u64))) +svuint32_t svsubhnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_n_u16))) +svuint8_t svsubhnt_n_u16(svuint8_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_n_s32))) +svint16_t svsubhnt_n_s32(svint16_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_n_s64))) +svint32_t svsubhnt_n_s64(svint32_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_n_s16))) +svint8_t svsubhnt_n_s16(svint8_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_u32))) +svuint16_t svsubhnt_u32(svuint16_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_u64))) +svuint32_t svsubhnt_u64(svuint32_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_u16))) +svuint8_t svsubhnt_u16(svuint8_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_s32))) +svint16_t svsubhnt_s32(svint16_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_s64))) +svint32_t svsubhnt_s64(svint32_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_s16))) +svint8_t svsubhnt_s16(svint8_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_n_s32))) +svint32_t svsublb_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_n_s64))) +svint64_t svsublb_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_n_s16))) +svint16_t svsublb_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_n_u32))) +svuint32_t svsublb_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_n_u64))) +svuint64_t svsublb_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_n_u16))) +svuint16_t svsublb_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_s32))) +svint32_t svsublb_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_s64))) +svint64_t svsublb_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_s16))) +svint16_t svsublb_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_u32))) +svuint32_t svsublb_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_u64))) +svuint64_t svsublb_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_u16))) +svuint16_t svsublb_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublbt_n_s32))) +svint32_t svsublbt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublbt_n_s64))) +svint64_t svsublbt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublbt_n_s16))) +svint16_t svsublbt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublbt_s32))) +svint32_t svsublbt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublbt_s64))) +svint64_t svsublbt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublbt_s16))) +svint16_t svsublbt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_n_s32))) +svint32_t svsublt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_n_s64))) +svint64_t svsublt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_n_s16))) +svint16_t svsublt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_n_u32))) +svuint32_t svsublt_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_n_u64))) +svuint64_t svsublt_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_n_u16))) +svuint16_t svsublt_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_s32))) +svint32_t svsublt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_s64))) +svint64_t svsublt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_s16))) +svint16_t svsublt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_u32))) +svuint32_t svsublt_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_u64))) +svuint64_t svsublt_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_u16))) +svuint16_t svsublt_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubltb_n_s32))) +svint32_t svsubltb_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubltb_n_s64))) +svint64_t svsubltb_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubltb_n_s16))) +svint16_t svsubltb_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubltb_s32))) +svint32_t svsubltb_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubltb_s64))) +svint64_t svsubltb_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubltb_s16))) +svint16_t svsubltb_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_n_s32))) +svint32_t svsubwb_n_s32(svint32_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_n_s64))) +svint64_t svsubwb_n_s64(svint64_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_n_s16))) +svint16_t svsubwb_n_s16(svint16_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_n_u32))) +svuint32_t svsubwb_n_u32(svuint32_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_n_u64))) +svuint64_t svsubwb_n_u64(svuint64_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_n_u16))) +svuint16_t svsubwb_n_u16(svuint16_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_s32))) +svint32_t svsubwb_s32(svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_s64))) +svint64_t svsubwb_s64(svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_s16))) +svint16_t svsubwb_s16(svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_u32))) +svuint32_t svsubwb_u32(svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_u64))) +svuint64_t svsubwb_u64(svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_u16))) +svuint16_t svsubwb_u16(svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_n_s32))) +svint32_t svsubwt_n_s32(svint32_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_n_s64))) +svint64_t svsubwt_n_s64(svint64_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_n_s16))) +svint16_t svsubwt_n_s16(svint16_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_n_u32))) +svuint32_t svsubwt_n_u32(svuint32_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_n_u64))) +svuint64_t svsubwt_n_u64(svuint64_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_n_u16))) +svuint16_t svsubwt_n_u16(svuint16_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_s32))) +svint32_t svsubwt_s32(svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_s64))) +svint64_t svsubwt_s64(svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_s16))) +svint16_t svsubwt_s16(svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_u32))) +svuint32_t svsubwt_u32(svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_u64))) +svuint64_t svsubwt_u64(svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_u16))) +svuint16_t svsubwt_u16(svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_u8))) +svuint8_t svtbl2_u8(svuint8x2_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_u32))) +svuint32_t svtbl2_u32(svuint32x2_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_u64))) +svuint64_t svtbl2_u64(svuint64x2_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_u16))) +svuint16_t svtbl2_u16(svuint16x2_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_s8))) +svint8_t svtbl2_s8(svint8x2_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_f64))) +svfloat64_t svtbl2_f64(svfloat64x2_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_f32))) +svfloat32_t svtbl2_f32(svfloat32x2_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_f16))) +svfloat16_t svtbl2_f16(svfloat16x2_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_s32))) +svint32_t svtbl2_s32(svint32x2_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_s64))) +svint64_t svtbl2_s64(svint64x2_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_s16))) +svint16_t svtbl2_s16(svint16x2_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_u8))) +svuint8_t svtbx_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_u32))) +svuint32_t svtbx_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_u64))) +svuint64_t svtbx_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_u16))) +svuint16_t svtbx_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_s8))) +svint8_t svtbx_s8(svint8_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_f64))) +svfloat64_t svtbx_f64(svfloat64_t, svfloat64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_f32))) +svfloat32_t svtbx_f32(svfloat32_t, svfloat32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_f16))) +svfloat16_t svtbx_f16(svfloat16_t, svfloat16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_s32))) +svint32_t svtbx_s32(svint32_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_s64))) +svint64_t svtbx_s64(svint64_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_s16))) +svint16_t svtbx_s16(svint16_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s8_m))) +svint8_t svuqadd_n_s8_m(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s32_m))) +svint32_t svuqadd_n_s32_m(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s64_m))) +svint64_t svuqadd_n_s64_m(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s16_m))) +svint16_t svuqadd_n_s16_m(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s8_x))) +svint8_t svuqadd_n_s8_x(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s32_x))) +svint32_t svuqadd_n_s32_x(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s64_x))) +svint64_t svuqadd_n_s64_x(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s16_x))) +svint16_t svuqadd_n_s16_x(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s8_z))) +svint8_t svuqadd_n_s8_z(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s32_z))) +svint32_t svuqadd_n_s32_z(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s64_z))) +svint64_t svuqadd_n_s64_z(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s16_z))) +svint16_t svuqadd_n_s16_z(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s8_m))) +svint8_t svuqadd_s8_m(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s32_m))) +svint32_t svuqadd_s32_m(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s64_m))) +svint64_t svuqadd_s64_m(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s16_m))) +svint16_t svuqadd_s16_m(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s8_x))) +svint8_t svuqadd_s8_x(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s32_x))) +svint32_t svuqadd_s32_x(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s64_x))) +svint64_t svuqadd_s64_x(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s16_x))) +svint16_t svuqadd_s16_x(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s8_z))) +svint8_t svuqadd_s8_z(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s32_z))) +svint32_t svuqadd_s32_z(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s64_z))) +svint64_t svuqadd_s64_z(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s16_z))) +svint16_t svuqadd_s16_z(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b8_s32))) +svbool_t svwhilege_b8_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b32_s32))) +svbool_t svwhilege_b32_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b64_s32))) +svbool_t svwhilege_b64_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b16_s32))) +svbool_t svwhilege_b16_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b8_s64))) +svbool_t svwhilege_b8_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b32_s64))) +svbool_t svwhilege_b32_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b64_s64))) +svbool_t svwhilege_b64_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b16_s64))) +svbool_t svwhilege_b16_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b8_u32))) +svbool_t svwhilege_b8_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b32_u32))) +svbool_t svwhilege_b32_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b64_u32))) +svbool_t svwhilege_b64_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b16_u32))) +svbool_t svwhilege_b16_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b8_u64))) +svbool_t svwhilege_b8_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b32_u64))) +svbool_t svwhilege_b32_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b64_u64))) +svbool_t svwhilege_b64_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b16_u64))) +svbool_t svwhilege_b16_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b8_s32))) +svbool_t svwhilegt_b8_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b32_s32))) +svbool_t svwhilegt_b32_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b64_s32))) +svbool_t svwhilegt_b64_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b16_s32))) +svbool_t svwhilegt_b16_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b8_s64))) +svbool_t svwhilegt_b8_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b32_s64))) +svbool_t svwhilegt_b32_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b64_s64))) +svbool_t svwhilegt_b64_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b16_s64))) +svbool_t svwhilegt_b16_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b8_u32))) +svbool_t svwhilegt_b8_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b32_u32))) +svbool_t svwhilegt_b32_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b64_u32))) +svbool_t svwhilegt_b64_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b16_u32))) +svbool_t svwhilegt_b16_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b8_u64))) +svbool_t svwhilegt_b8_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b32_u64))) +svbool_t svwhilegt_b32_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b64_u64))) +svbool_t svwhilegt_b64_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b16_u64))) +svbool_t svwhilegt_b16_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_u8))) +svbool_t svwhilerw_u8(uint8_t const *, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_s8))) +svbool_t svwhilerw_s8(int8_t const *, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_u64))) +svbool_t svwhilerw_u64(uint64_t const *, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_f64))) +svbool_t svwhilerw_f64(float64_t const *, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_s64))) +svbool_t svwhilerw_s64(int64_t const *, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_u16))) +svbool_t svwhilerw_u16(uint16_t const *, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_f16))) +svbool_t svwhilerw_f16(float16_t const *, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_s16))) +svbool_t svwhilerw_s16(int16_t const *, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_u32))) +svbool_t svwhilerw_u32(uint32_t const *, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_f32))) +svbool_t svwhilerw_f32(float32_t const *, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_s32))) +svbool_t svwhilerw_s32(int32_t const *, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_u8))) +svbool_t svwhilewr_u8(uint8_t const *, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_s8))) +svbool_t svwhilewr_s8(int8_t const *, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_u64))) +svbool_t svwhilewr_u64(uint64_t const *, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_f64))) +svbool_t svwhilewr_f64(float64_t const *, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_s64))) +svbool_t svwhilewr_s64(int64_t const *, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_u16))) +svbool_t svwhilewr_u16(uint16_t const *, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_f16))) +svbool_t svwhilewr_f16(float16_t const *, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_s16))) +svbool_t svwhilewr_s16(int16_t const *, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_u32))) +svbool_t svwhilewr_u32(uint32_t const *, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_f32))) +svbool_t svwhilewr_f32(float32_t const *, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_s32))) +svbool_t svwhilewr_s32(int32_t const *, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_u8))) +svuint8_t svxar_n_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_u32))) +svuint32_t svxar_n_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_u64))) +svuint64_t svxar_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_u16))) +svuint16_t svxar_n_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_s8))) +svint8_t svxar_n_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_s32))) +svint32_t svxar_n_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_s64))) +svint64_t svxar_n_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_s16))) +svint16_t svxar_n_s16(svint16_t, svint16_t, uint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_s8))) svint8_t svaba(svint8_t, svint8_t, int8_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_s32))) @@ -17527,8 +23230,10 @@ svint16_t svxar(svint16_t, svint16_t, uint64_t); #endif //defined(__ARM_FEATURE_SVE2) #if defined(__ARM_FEATURE_SVE2) && defined(__ARM_FEATURE_BF16_SCALAR_ARITHMETIC) -#define svwhilerw_bf16(...) __builtin_sve_svwhilerw_bf16(__VA_ARGS__) -#define svwhilewr_bf16(...) __builtin_sve_svwhilewr_bf16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_bf16))) +svbool_t svwhilerw_bf16(bfloat16_t const *, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_bf16))) +svbool_t svwhilewr_bf16(bfloat16_t const *, bfloat16_t const *); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_bf16))) svbool_t svwhilerw(bfloat16_t const *, bfloat16_t const *); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_bf16))) @@ -17536,8 +23241,10 @@ svbool_t svwhilewr(bfloat16_t const *, bfloat16_t const *); #endif //defined(__ARM_FEATURE_SVE2) && defined(__ARM_FEATURE_BF16_SCALAR_ARITHMETIC) #if defined(__ARM_FEATURE_SVE2) && defined(__ARM_FEATURE_SVE_BF16) -#define svtbl2_bf16(...) __builtin_sve_svtbl2_bf16(__VA_ARGS__) -#define svtbx_bf16(...) __builtin_sve_svtbx_bf16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_bf16))) +svbfloat16_t svtbl2_bf16(svbfloat16x2_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_bf16))) +svbfloat16_t svtbx_bf16(svbfloat16_t, svbfloat16_t, svuint16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_bf16))) svbfloat16_t svtbl2(svbfloat16x2_t, svuint16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_bf16))) @@ -17545,14 +23252,22 @@ svbfloat16_t svtbx(svbfloat16_t, svbfloat16_t, svuint16_t); #endif //defined(__ARM_FEATURE_SVE2) && defined(__ARM_FEATURE_SVE_BF16) #if defined(__ARM_FEATURE_SVE2_AES) -#define svaesd_u8(...) __builtin_sve_svaesd_u8(__VA_ARGS__) -#define svaese_u8(...) __builtin_sve_svaese_u8(__VA_ARGS__) -#define svaesimc_u8(...) __builtin_sve_svaesimc_u8(__VA_ARGS__) -#define svaesmc_u8(...) __builtin_sve_svaesmc_u8(__VA_ARGS__) -#define svpmullb_pair_n_u64(...) __builtin_sve_svpmullb_pair_n_u64(__VA_ARGS__) -#define svpmullb_pair_u64(...) __builtin_sve_svpmullb_pair_u64(__VA_ARGS__) -#define svpmullt_pair_n_u64(...) __builtin_sve_svpmullt_pair_n_u64(__VA_ARGS__) -#define svpmullt_pair_u64(...) __builtin_sve_svpmullt_pair_u64(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaesd_u8))) +svuint8_t svaesd_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaese_u8))) +svuint8_t svaese_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaesimc_u8))) +svuint8_t svaesimc_u8(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaesmc_u8))) +svuint8_t svaesmc_u8(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_pair_n_u64))) +svuint64_t svpmullb_pair_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_pair_u64))) +svuint64_t svpmullb_pair_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_pair_n_u64))) +svuint64_t svpmullt_pair_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_pair_u64))) +svuint64_t svpmullt_pair_u64(svuint64_t, svuint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaesd_u8))) svuint8_t svaesd(svuint8_t, svuint8_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaese_u8))) @@ -17572,8 +23287,10 @@ svuint64_t svpmullt_pair(svuint64_t, svuint64_t); #endif //defined(__ARM_FEATURE_SVE2_AES) #if defined(__ARM_FEATURE_SVE2_SHA3) -#define svrax1_u64(...) __builtin_sve_svrax1_u64(__VA_ARGS__) -#define svrax1_s64(...) __builtin_sve_svrax1_s64(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrax1_u64))) +svuint64_t svrax1_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrax1_s64))) +svint64_t svrax1_s64(svint64_t, svint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrax1_u64))) svuint64_t svrax1(svuint64_t, svuint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrax1_s64))) @@ -17581,8 +23298,10 @@ svint64_t svrax1(svint64_t, svint64_t); #endif //defined(__ARM_FEATURE_SVE2_SHA3) #if defined(__ARM_FEATURE_SVE2_SM4) -#define svsm4e_u32(...) __builtin_sve_svsm4e_u32(__VA_ARGS__) -#define svsm4ekey_u32(...) __builtin_sve_svsm4ekey_u32(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsm4e_u32))) +svuint32_t svsm4e_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsm4ekey_u32))) +svuint32_t svsm4ekey_u32(svuint32_t, svuint32_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsm4e_u32))) svuint32_t svsm4e(svuint32_t, svuint32_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsm4ekey_u32))) @@ -17590,87 +23309,168 @@ svuint32_t svsm4ekey(svuint32_t, svuint32_t); #endif //defined(__ARM_FEATURE_SVE2_SM4) #if defined(__ARM_FEATURE_SVE_BF16) -#define svbfdot_n_f32(...) __builtin_sve_svbfdot_n_f32(__VA_ARGS__) -#define svbfdot_f32(...) __builtin_sve_svbfdot_f32(__VA_ARGS__) -#define svbfdot_lane_f32(...) __builtin_sve_svbfdot_lane_f32(__VA_ARGS__) -#define svbfmlalb_n_f32(...) __builtin_sve_svbfmlalb_n_f32(__VA_ARGS__) -#define svbfmlalb_f32(...) __builtin_sve_svbfmlalb_f32(__VA_ARGS__) -#define svbfmlalb_lane_f32(...) __builtin_sve_svbfmlalb_lane_f32(__VA_ARGS__) -#define svbfmlalt_n_f32(...) __builtin_sve_svbfmlalt_n_f32(__VA_ARGS__) -#define svbfmlalt_f32(...) __builtin_sve_svbfmlalt_f32(__VA_ARGS__) -#define svbfmlalt_lane_f32(...) __builtin_sve_svbfmlalt_lane_f32(__VA_ARGS__) -#define svbfmmla_f32(...) __builtin_sve_svbfmmla_f32(__VA_ARGS__) -#define svclasta_n_bf16(...) __builtin_sve_svclasta_n_bf16(__VA_ARGS__) -#define svclasta_bf16(...) __builtin_sve_svclasta_bf16(__VA_ARGS__) -#define svclastb_n_bf16(...) __builtin_sve_svclastb_n_bf16(__VA_ARGS__) -#define svclastb_bf16(...) __builtin_sve_svclastb_bf16(__VA_ARGS__) -#define svcnt_bf16_m(...) __builtin_sve_svcnt_bf16_m(__VA_ARGS__) -#define svcnt_bf16_x(...) __builtin_sve_svcnt_bf16_x(__VA_ARGS__) -#define svcnt_bf16_z(...) __builtin_sve_svcnt_bf16_z(__VA_ARGS__) -#define svcreate2_bf16(...) __builtin_sve_svcreate2_bf16(__VA_ARGS__) -#define svcreate3_bf16(...) __builtin_sve_svcreate3_bf16(__VA_ARGS__) -#define svcreate4_bf16(...) __builtin_sve_svcreate4_bf16(__VA_ARGS__) -#define svcvt_bf16_f32_m(...) __builtin_sve_svcvt_bf16_f32_m(__VA_ARGS__) -#define svcvt_bf16_f32_x(...) __builtin_sve_svcvt_bf16_f32_x(__VA_ARGS__) -#define svcvt_bf16_f32_z(...) __builtin_sve_svcvt_bf16_f32_z(__VA_ARGS__) -#define svcvtnt_bf16_f32_m(...) __builtin_sve_svcvtnt_bf16_f32_m(__VA_ARGS__) -#define svdup_n_bf16(...) __builtin_sve_svdup_n_bf16(__VA_ARGS__) -#define svdup_n_bf16_m(...) __builtin_sve_svdup_n_bf16_m(__VA_ARGS__) -#define svdup_n_bf16_x(...) __builtin_sve_svdup_n_bf16_x(__VA_ARGS__) -#define svdup_n_bf16_z(...) __builtin_sve_svdup_n_bf16_z(__VA_ARGS__) -#define svdup_lane_bf16(...) __builtin_sve_svdup_lane_bf16(__VA_ARGS__) -#define svdupq_n_bf16(...) __builtin_sve_svdupq_n_bf16(__VA_ARGS__) -#define svdupq_lane_bf16(...) __builtin_sve_svdupq_lane_bf16(__VA_ARGS__) -#define svext_bf16(...) __builtin_sve_svext_bf16(__VA_ARGS__) -#define svget2_bf16(...) __builtin_sve_svget2_bf16(__VA_ARGS__) -#define svget3_bf16(...) __builtin_sve_svget3_bf16(__VA_ARGS__) -#define svget4_bf16(...) __builtin_sve_svget4_bf16(__VA_ARGS__) -#define svinsr_n_bf16(...) __builtin_sve_svinsr_n_bf16(__VA_ARGS__) -#define svlasta_bf16(...) __builtin_sve_svlasta_bf16(__VA_ARGS__) -#define svlastb_bf16(...) __builtin_sve_svlastb_bf16(__VA_ARGS__) -#define svld1_bf16(...) __builtin_sve_svld1_bf16(__VA_ARGS__) -#define svld1_vnum_bf16(...) __builtin_sve_svld1_vnum_bf16(__VA_ARGS__) -#define svld1rq_bf16(...) __builtin_sve_svld1rq_bf16(__VA_ARGS__) -#define svld2_bf16(...) __builtin_sve_svld2_bf16(__VA_ARGS__) -#define svld2_vnum_bf16(...) __builtin_sve_svld2_vnum_bf16(__VA_ARGS__) -#define svld3_bf16(...) __builtin_sve_svld3_bf16(__VA_ARGS__) -#define svld3_vnum_bf16(...) __builtin_sve_svld3_vnum_bf16(__VA_ARGS__) -#define svld4_bf16(...) __builtin_sve_svld4_bf16(__VA_ARGS__) -#define svld4_vnum_bf16(...) __builtin_sve_svld4_vnum_bf16(__VA_ARGS__) -#define svldff1_bf16(...) __builtin_sve_svldff1_bf16(__VA_ARGS__) -#define svldff1_vnum_bf16(...) __builtin_sve_svldff1_vnum_bf16(__VA_ARGS__) -#define svldnf1_bf16(...) __builtin_sve_svldnf1_bf16(__VA_ARGS__) -#define svldnf1_vnum_bf16(...) __builtin_sve_svldnf1_vnum_bf16(__VA_ARGS__) -#define svldnt1_bf16(...) __builtin_sve_svldnt1_bf16(__VA_ARGS__) -#define svldnt1_vnum_bf16(...) __builtin_sve_svldnt1_vnum_bf16(__VA_ARGS__) -#define svlen_bf16(...) __builtin_sve_svlen_bf16(__VA_ARGS__) -#define svrev_bf16(...) __builtin_sve_svrev_bf16(__VA_ARGS__) -#define svsel_bf16(...) __builtin_sve_svsel_bf16(__VA_ARGS__) -#define svset2_bf16(...) __builtin_sve_svset2_bf16(__VA_ARGS__) -#define svset3_bf16(...) __builtin_sve_svset3_bf16(__VA_ARGS__) -#define svset4_bf16(...) __builtin_sve_svset4_bf16(__VA_ARGS__) -#define svsplice_bf16(...) __builtin_sve_svsplice_bf16(__VA_ARGS__) -#define svst1_bf16(...) __builtin_sve_svst1_bf16(__VA_ARGS__) -#define svst1_vnum_bf16(...) __builtin_sve_svst1_vnum_bf16(__VA_ARGS__) -#define svst2_bf16(...) __builtin_sve_svst2_bf16(__VA_ARGS__) -#define svst2_vnum_bf16(...) __builtin_sve_svst2_vnum_bf16(__VA_ARGS__) -#define svst3_bf16(...) __builtin_sve_svst3_bf16(__VA_ARGS__) -#define svst3_vnum_bf16(...) __builtin_sve_svst3_vnum_bf16(__VA_ARGS__) -#define svst4_bf16(...) __builtin_sve_svst4_bf16(__VA_ARGS__) -#define svst4_vnum_bf16(...) __builtin_sve_svst4_vnum_bf16(__VA_ARGS__) -#define svstnt1_bf16(...) __builtin_sve_svstnt1_bf16(__VA_ARGS__) -#define svstnt1_vnum_bf16(...) __builtin_sve_svstnt1_vnum_bf16(__VA_ARGS__) -#define svtbl_bf16(...) __builtin_sve_svtbl_bf16(__VA_ARGS__) -#define svtrn1_bf16(...) __builtin_sve_svtrn1_bf16(__VA_ARGS__) -#define svtrn2_bf16(...) __builtin_sve_svtrn2_bf16(__VA_ARGS__) -#define svundef2_bf16(...) __builtin_sve_svundef2_bf16(__VA_ARGS__) -#define svundef3_bf16(...) __builtin_sve_svundef3_bf16(__VA_ARGS__) -#define svundef4_bf16(...) __builtin_sve_svundef4_bf16(__VA_ARGS__) -#define svundef_bf16(...) __builtin_sve_svundef_bf16(__VA_ARGS__) -#define svuzp1_bf16(...) __builtin_sve_svuzp1_bf16(__VA_ARGS__) -#define svuzp2_bf16(...) __builtin_sve_svuzp2_bf16(__VA_ARGS__) -#define svzip1_bf16(...) __builtin_sve_svzip1_bf16(__VA_ARGS__) -#define svzip2_bf16(...) __builtin_sve_svzip2_bf16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfdot_n_f32))) +svfloat32_t svbfdot_n_f32(svfloat32_t, svbfloat16_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfdot_f32))) +svfloat32_t svbfdot_f32(svfloat32_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfdot_lane_f32))) +svfloat32_t svbfdot_lane_f32(svfloat32_t, svbfloat16_t, svbfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmlalb_n_f32))) +svfloat32_t svbfmlalb_n_f32(svfloat32_t, svbfloat16_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmlalb_f32))) +svfloat32_t svbfmlalb_f32(svfloat32_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmlalb_lane_f32))) +svfloat32_t svbfmlalb_lane_f32(svfloat32_t, svbfloat16_t, svbfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmlalt_n_f32))) +svfloat32_t svbfmlalt_n_f32(svfloat32_t, svbfloat16_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmlalt_f32))) +svfloat32_t svbfmlalt_f32(svfloat32_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmlalt_lane_f32))) +svfloat32_t svbfmlalt_lane_f32(svfloat32_t, svbfloat16_t, svbfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmmla_f32))) +svfloat32_t svbfmmla_f32(svfloat32_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_bf16))) +bfloat16_t svclasta_n_bf16(svbool_t, bfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_bf16))) +svbfloat16_t svclasta_bf16(svbool_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_bf16))) +bfloat16_t svclastb_n_bf16(svbool_t, bfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_bf16))) +svbfloat16_t svclastb_bf16(svbool_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_bf16_m))) +svuint16_t svcnt_bf16_m(svuint16_t, svbool_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_bf16_x))) +svuint16_t svcnt_bf16_x(svbool_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_bf16_z))) +svuint16_t svcnt_bf16_z(svbool_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_bf16))) +svbfloat16x2_t svcreate2_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_bf16))) +svbfloat16x3_t svcreate3_bf16(svbfloat16_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_bf16))) +svbfloat16x4_t svcreate4_bf16(svbfloat16_t, svbfloat16_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_bf16_f32_m))) +svbfloat16_t svcvt_bf16_f32_m(svbfloat16_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_bf16_f32_x))) +svbfloat16_t svcvt_bf16_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_bf16_f32_z))) +svbfloat16_t svcvt_bf16_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtnt_bf16_f32_m))) +svbfloat16_t svcvtnt_bf16_f32_m(svbfloat16_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_bf16))) +svbfloat16_t svdup_n_bf16(bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_bf16_m))) +svbfloat16_t svdup_n_bf16_m(svbfloat16_t, svbool_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_bf16_x))) +svbfloat16_t svdup_n_bf16_x(svbool_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_bf16_z))) +svbfloat16_t svdup_n_bf16_z(svbool_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_bf16))) +svbfloat16_t svdup_lane_bf16(svbfloat16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_bf16))) +svbfloat16_t svdupq_n_bf16(bfloat16_t, bfloat16_t, bfloat16_t, bfloat16_t, bfloat16_t, bfloat16_t, bfloat16_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_bf16))) +svbfloat16_t svdupq_lane_bf16(svbfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_bf16))) +svbfloat16_t svext_bf16(svbfloat16_t, svbfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_bf16))) +svbfloat16_t svget2_bf16(svbfloat16x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_bf16))) +svbfloat16_t svget3_bf16(svbfloat16x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_bf16))) +svbfloat16_t svget4_bf16(svbfloat16x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_bf16))) +svbfloat16_t svinsr_n_bf16(svbfloat16_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_bf16))) +bfloat16_t svlasta_bf16(svbool_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_bf16))) +bfloat16_t svlastb_bf16(svbool_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_bf16))) +svbfloat16_t svld1_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_bf16))) +svbfloat16_t svld1_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_bf16))) +svbfloat16_t svld1rq_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_bf16))) +svbfloat16x2_t svld2_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_bf16))) +svbfloat16x2_t svld2_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_bf16))) +svbfloat16x3_t svld3_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_bf16))) +svbfloat16x3_t svld3_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_bf16))) +svbfloat16x4_t svld4_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_bf16))) +svbfloat16x4_t svld4_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_bf16))) +svbfloat16_t svldff1_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_bf16))) +svbfloat16_t svldff1_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_bf16))) +svbfloat16_t svldnf1_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_bf16))) +svbfloat16_t svldnf1_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_bf16))) +svbfloat16_t svldnt1_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_bf16))) +svbfloat16_t svldnt1_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_bf16))) +uint64_t svlen_bf16(svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_bf16))) +svbfloat16_t svrev_bf16(svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_bf16))) +svbfloat16_t svsel_bf16(svbool_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_bf16))) +svbfloat16x2_t svset2_bf16(svbfloat16x2_t, uint64_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_bf16))) +svbfloat16x3_t svset3_bf16(svbfloat16x3_t, uint64_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_bf16))) +svbfloat16x4_t svset4_bf16(svbfloat16x4_t, uint64_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_bf16))) +svbfloat16_t svsplice_bf16(svbool_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_bf16))) +void svst1_bf16(svbool_t, bfloat16_t *, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_bf16))) +void svst1_vnum_bf16(svbool_t, bfloat16_t *, int64_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_bf16))) +void svst2_bf16(svbool_t, bfloat16_t *, svbfloat16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_bf16))) +void svst2_vnum_bf16(svbool_t, bfloat16_t *, int64_t, svbfloat16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_bf16))) +void svst3_bf16(svbool_t, bfloat16_t *, svbfloat16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_bf16))) +void svst3_vnum_bf16(svbool_t, bfloat16_t *, int64_t, svbfloat16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_bf16))) +void svst4_bf16(svbool_t, bfloat16_t *, svbfloat16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_bf16))) +void svst4_vnum_bf16(svbool_t, bfloat16_t *, int64_t, svbfloat16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_bf16))) +void svstnt1_bf16(svbool_t, bfloat16_t *, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_bf16))) +void svstnt1_vnum_bf16(svbool_t, bfloat16_t *, int64_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_bf16))) +svbfloat16_t svtbl_bf16(svbfloat16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_bf16))) +svbfloat16_t svtrn1_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_bf16))) +svbfloat16_t svtrn2_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_bf16))) +svbfloat16x2_t svundef2_bf16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_bf16))) +svbfloat16x3_t svundef3_bf16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_bf16))) +svbfloat16x4_t svundef4_bf16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_bf16))) +svbfloat16_t svundef_bf16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_bf16))) +svbfloat16_t svuzp1_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_bf16))) +svbfloat16_t svuzp2_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_bf16))) +svbfloat16_t svzip1_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_bf16))) +svbfloat16_t svzip2_bf16(svbfloat16_t, svbfloat16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfdot_n_f32))) svfloat32_t svbfdot(svfloat32_t, svbfloat16_t, bfloat16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfdot_f32))) @@ -17828,90 +23628,169 @@ svbfloat16_t svzip2(svbfloat16_t, svbfloat16_t); #endif //defined(__ARM_FEATURE_SVE_BF16) #if defined(__ARM_FEATURE_SVE_MATMUL_FP32) -#define svmmla_f32(...) __builtin_sve_svmmla_f32(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_f32))) +svfloat32_t svmmla_f32(svfloat32_t, svfloat32_t, svfloat32_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_f32))) svfloat32_t svmmla(svfloat32_t, svfloat32_t, svfloat32_t); #endif //defined(__ARM_FEATURE_SVE_MATMUL_FP32) #if defined(__ARM_FEATURE_SVE_MATMUL_FP64) -#define svld1ro_u8(...) __builtin_sve_svld1ro_u8(__VA_ARGS__) -#define svld1ro_u32(...) __builtin_sve_svld1ro_u32(__VA_ARGS__) -#define svld1ro_u64(...) __builtin_sve_svld1ro_u64(__VA_ARGS__) -#define svld1ro_u16(...) __builtin_sve_svld1ro_u16(__VA_ARGS__) -#define svld1ro_s8(...) __builtin_sve_svld1ro_s8(__VA_ARGS__) -#define svld1ro_f64(...) __builtin_sve_svld1ro_f64(__VA_ARGS__) -#define svld1ro_f32(...) __builtin_sve_svld1ro_f32(__VA_ARGS__) -#define svld1ro_f16(...) __builtin_sve_svld1ro_f16(__VA_ARGS__) -#define svld1ro_s32(...) __builtin_sve_svld1ro_s32(__VA_ARGS__) -#define svld1ro_s64(...) __builtin_sve_svld1ro_s64(__VA_ARGS__) -#define svld1ro_s16(...) __builtin_sve_svld1ro_s16(__VA_ARGS__) -#define svmmla_f64(...) __builtin_sve_svmmla_f64(__VA_ARGS__) -#define svtrn1q_u8(...) __builtin_sve_svtrn1q_u8(__VA_ARGS__) -#define svtrn1q_u32(...) __builtin_sve_svtrn1q_u32(__VA_ARGS__) -#define svtrn1q_u64(...) __builtin_sve_svtrn1q_u64(__VA_ARGS__) -#define svtrn1q_u16(...) __builtin_sve_svtrn1q_u16(__VA_ARGS__) -#define svtrn1q_s8(...) __builtin_sve_svtrn1q_s8(__VA_ARGS__) -#define svtrn1q_f64(...) __builtin_sve_svtrn1q_f64(__VA_ARGS__) -#define svtrn1q_f32(...) __builtin_sve_svtrn1q_f32(__VA_ARGS__) -#define svtrn1q_f16(...) __builtin_sve_svtrn1q_f16(__VA_ARGS__) -#define svtrn1q_s32(...) __builtin_sve_svtrn1q_s32(__VA_ARGS__) -#define svtrn1q_s64(...) __builtin_sve_svtrn1q_s64(__VA_ARGS__) -#define svtrn1q_s16(...) __builtin_sve_svtrn1q_s16(__VA_ARGS__) -#define svtrn2q_u8(...) __builtin_sve_svtrn2q_u8(__VA_ARGS__) -#define svtrn2q_u32(...) __builtin_sve_svtrn2q_u32(__VA_ARGS__) -#define svtrn2q_u64(...) __builtin_sve_svtrn2q_u64(__VA_ARGS__) -#define svtrn2q_u16(...) __builtin_sve_svtrn2q_u16(__VA_ARGS__) -#define svtrn2q_s8(...) __builtin_sve_svtrn2q_s8(__VA_ARGS__) -#define svtrn2q_f64(...) __builtin_sve_svtrn2q_f64(__VA_ARGS__) -#define svtrn2q_f32(...) __builtin_sve_svtrn2q_f32(__VA_ARGS__) -#define svtrn2q_f16(...) __builtin_sve_svtrn2q_f16(__VA_ARGS__) -#define svtrn2q_s32(...) __builtin_sve_svtrn2q_s32(__VA_ARGS__) -#define svtrn2q_s64(...) __builtin_sve_svtrn2q_s64(__VA_ARGS__) -#define svtrn2q_s16(...) __builtin_sve_svtrn2q_s16(__VA_ARGS__) -#define svuzp1q_u8(...) __builtin_sve_svuzp1q_u8(__VA_ARGS__) -#define svuzp1q_u32(...) __builtin_sve_svuzp1q_u32(__VA_ARGS__) -#define svuzp1q_u64(...) __builtin_sve_svuzp1q_u64(__VA_ARGS__) -#define svuzp1q_u16(...) __builtin_sve_svuzp1q_u16(__VA_ARGS__) -#define svuzp1q_s8(...) __builtin_sve_svuzp1q_s8(__VA_ARGS__) -#define svuzp1q_f64(...) __builtin_sve_svuzp1q_f64(__VA_ARGS__) -#define svuzp1q_f32(...) __builtin_sve_svuzp1q_f32(__VA_ARGS__) -#define svuzp1q_f16(...) __builtin_sve_svuzp1q_f16(__VA_ARGS__) -#define svuzp1q_s32(...) __builtin_sve_svuzp1q_s32(__VA_ARGS__) -#define svuzp1q_s64(...) __builtin_sve_svuzp1q_s64(__VA_ARGS__) -#define svuzp1q_s16(...) __builtin_sve_svuzp1q_s16(__VA_ARGS__) -#define svuzp2q_u8(...) __builtin_sve_svuzp2q_u8(__VA_ARGS__) -#define svuzp2q_u32(...) __builtin_sve_svuzp2q_u32(__VA_ARGS__) -#define svuzp2q_u64(...) __builtin_sve_svuzp2q_u64(__VA_ARGS__) -#define svuzp2q_u16(...) __builtin_sve_svuzp2q_u16(__VA_ARGS__) -#define svuzp2q_s8(...) __builtin_sve_svuzp2q_s8(__VA_ARGS__) -#define svuzp2q_f64(...) __builtin_sve_svuzp2q_f64(__VA_ARGS__) -#define svuzp2q_f32(...) __builtin_sve_svuzp2q_f32(__VA_ARGS__) -#define svuzp2q_f16(...) __builtin_sve_svuzp2q_f16(__VA_ARGS__) -#define svuzp2q_s32(...) __builtin_sve_svuzp2q_s32(__VA_ARGS__) -#define svuzp2q_s64(...) __builtin_sve_svuzp2q_s64(__VA_ARGS__) -#define svuzp2q_s16(...) __builtin_sve_svuzp2q_s16(__VA_ARGS__) -#define svzip1q_u8(...) __builtin_sve_svzip1q_u8(__VA_ARGS__) -#define svzip1q_u32(...) __builtin_sve_svzip1q_u32(__VA_ARGS__) -#define svzip1q_u64(...) __builtin_sve_svzip1q_u64(__VA_ARGS__) -#define svzip1q_u16(...) __builtin_sve_svzip1q_u16(__VA_ARGS__) -#define svzip1q_s8(...) __builtin_sve_svzip1q_s8(__VA_ARGS__) -#define svzip1q_f64(...) __builtin_sve_svzip1q_f64(__VA_ARGS__) -#define svzip1q_f32(...) __builtin_sve_svzip1q_f32(__VA_ARGS__) -#define svzip1q_f16(...) __builtin_sve_svzip1q_f16(__VA_ARGS__) -#define svzip1q_s32(...) __builtin_sve_svzip1q_s32(__VA_ARGS__) -#define svzip1q_s64(...) __builtin_sve_svzip1q_s64(__VA_ARGS__) -#define svzip1q_s16(...) __builtin_sve_svzip1q_s16(__VA_ARGS__) -#define svzip2q_u8(...) __builtin_sve_svzip2q_u8(__VA_ARGS__) -#define svzip2q_u32(...) __builtin_sve_svzip2q_u32(__VA_ARGS__) -#define svzip2q_u64(...) __builtin_sve_svzip2q_u64(__VA_ARGS__) -#define svzip2q_u16(...) __builtin_sve_svzip2q_u16(__VA_ARGS__) -#define svzip2q_s8(...) __builtin_sve_svzip2q_s8(__VA_ARGS__) -#define svzip2q_f64(...) __builtin_sve_svzip2q_f64(__VA_ARGS__) -#define svzip2q_f32(...) __builtin_sve_svzip2q_f32(__VA_ARGS__) -#define svzip2q_f16(...) __builtin_sve_svzip2q_f16(__VA_ARGS__) -#define svzip2q_s32(...) __builtin_sve_svzip2q_s32(__VA_ARGS__) -#define svzip2q_s64(...) __builtin_sve_svzip2q_s64(__VA_ARGS__) -#define svzip2q_s16(...) __builtin_sve_svzip2q_s16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_u8))) +svuint8_t svld1ro_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_u32))) +svuint32_t svld1ro_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_u64))) +svuint64_t svld1ro_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_u16))) +svuint16_t svld1ro_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_s8))) +svint8_t svld1ro_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_f64))) +svfloat64_t svld1ro_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_f32))) +svfloat32_t svld1ro_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_f16))) +svfloat16_t svld1ro_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_s32))) +svint32_t svld1ro_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_s64))) +svint64_t svld1ro_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_s16))) +svint16_t svld1ro_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_f64))) +svfloat64_t svmmla_f64(svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_u8))) +svuint8_t svtrn1q_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_u32))) +svuint32_t svtrn1q_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_u64))) +svuint64_t svtrn1q_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_u16))) +svuint16_t svtrn1q_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_s8))) +svint8_t svtrn1q_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_f64))) +svfloat64_t svtrn1q_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_f32))) +svfloat32_t svtrn1q_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_f16))) +svfloat16_t svtrn1q_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_s32))) +svint32_t svtrn1q_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_s64))) +svint64_t svtrn1q_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_s16))) +svint16_t svtrn1q_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_u8))) +svuint8_t svtrn2q_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_u32))) +svuint32_t svtrn2q_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_u64))) +svuint64_t svtrn2q_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_u16))) +svuint16_t svtrn2q_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_s8))) +svint8_t svtrn2q_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_f64))) +svfloat64_t svtrn2q_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_f32))) +svfloat32_t svtrn2q_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_f16))) +svfloat16_t svtrn2q_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_s32))) +svint32_t svtrn2q_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_s64))) +svint64_t svtrn2q_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_s16))) +svint16_t svtrn2q_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_u8))) +svuint8_t svuzp1q_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_u32))) +svuint32_t svuzp1q_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_u64))) +svuint64_t svuzp1q_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_u16))) +svuint16_t svuzp1q_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_s8))) +svint8_t svuzp1q_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_f64))) +svfloat64_t svuzp1q_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_f32))) +svfloat32_t svuzp1q_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_f16))) +svfloat16_t svuzp1q_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_s32))) +svint32_t svuzp1q_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_s64))) +svint64_t svuzp1q_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_s16))) +svint16_t svuzp1q_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_u8))) +svuint8_t svuzp2q_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_u32))) +svuint32_t svuzp2q_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_u64))) +svuint64_t svuzp2q_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_u16))) +svuint16_t svuzp2q_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_s8))) +svint8_t svuzp2q_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_f64))) +svfloat64_t svuzp2q_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_f32))) +svfloat32_t svuzp2q_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_f16))) +svfloat16_t svuzp2q_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_s32))) +svint32_t svuzp2q_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_s64))) +svint64_t svuzp2q_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_s16))) +svint16_t svuzp2q_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_u8))) +svuint8_t svzip1q_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_u32))) +svuint32_t svzip1q_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_u64))) +svuint64_t svzip1q_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_u16))) +svuint16_t svzip1q_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_s8))) +svint8_t svzip1q_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_f64))) +svfloat64_t svzip1q_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_f32))) +svfloat32_t svzip1q_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_f16))) +svfloat16_t svzip1q_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_s32))) +svint32_t svzip1q_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_s64))) +svint64_t svzip1q_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_s16))) +svint16_t svzip1q_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_u8))) +svuint8_t svzip2q_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_u32))) +svuint32_t svzip2q_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_u64))) +svuint64_t svzip2q_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_u16))) +svuint16_t svzip2q_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_s8))) +svint8_t svzip2q_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_f64))) +svfloat64_t svzip2q_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_f32))) +svfloat32_t svzip2q_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_f16))) +svfloat16_t svzip2q_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_s32))) +svint32_t svzip2q_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_s64))) +svint64_t svzip2q_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_s16))) +svint16_t svzip2q_s16(svint16_t, svint16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_u8))) svuint8_t svld1ro(svbool_t, uint8_t const *); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_u32))) @@ -18071,13 +23950,20 @@ svint16_t svzip2q(svint16_t, svint16_t); #endif //defined(__ARM_FEATURE_SVE_MATMUL_FP64) #if defined(__ARM_FEATURE_SVE_MATMUL_FP64) && defined(__ARM_FEATURE_SVE_BF16) -#define svld1ro_bf16(...) __builtin_sve_svld1ro_bf16(__VA_ARGS__) -#define svtrn1q_bf16(...) __builtin_sve_svtrn1q_bf16(__VA_ARGS__) -#define svtrn2q_bf16(...) __builtin_sve_svtrn2q_bf16(__VA_ARGS__) -#define svuzp1q_bf16(...) __builtin_sve_svuzp1q_bf16(__VA_ARGS__) -#define svuzp2q_bf16(...) __builtin_sve_svuzp2q_bf16(__VA_ARGS__) -#define svzip1q_bf16(...) __builtin_sve_svzip1q_bf16(__VA_ARGS__) -#define svzip2q_bf16(...) __builtin_sve_svzip2q_bf16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_bf16))) +svbfloat16_t svld1ro_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_bf16))) +svbfloat16_t svtrn1q_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_bf16))) +svbfloat16_t svtrn2q_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_bf16))) +svbfloat16_t svuzp1q_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_bf16))) +svbfloat16_t svuzp2q_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_bf16))) +svbfloat16_t svzip1q_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_bf16))) +svbfloat16_t svzip2q_bf16(svbfloat16_t, svbfloat16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_bf16))) svbfloat16_t svld1ro(svbool_t, bfloat16_t const *); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_bf16))) @@ -18095,15 +23981,24 @@ svbfloat16_t svzip2q(svbfloat16_t, svbfloat16_t); #endif //defined(__ARM_FEATURE_SVE_MATMUL_FP64) && defined(__ARM_FEATURE_SVE_BF16) #if defined(__ARM_FEATURE_SVE_MATMUL_INT8) -#define svmmla_s32(...) __builtin_sve_svmmla_s32(__VA_ARGS__) -#define svmmla_u32(...) __builtin_sve_svmmla_u32(__VA_ARGS__) -#define svsudot_n_s32(...) __builtin_sve_svsudot_n_s32(__VA_ARGS__) -#define svsudot_s32(...) __builtin_sve_svsudot_s32(__VA_ARGS__) -#define svsudot_lane_s32(...) __builtin_sve_svsudot_lane_s32(__VA_ARGS__) -#define svusdot_n_s32(...) __builtin_sve_svusdot_n_s32(__VA_ARGS__) -#define svusdot_s32(...) __builtin_sve_svusdot_s32(__VA_ARGS__) -#define svusdot_lane_s32(...) __builtin_sve_svusdot_lane_s32(__VA_ARGS__) -#define svusmmla_s32(...) __builtin_sve_svusmmla_s32(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_s32))) +svint32_t svmmla_s32(svint32_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_u32))) +svuint32_t svmmla_u32(svuint32_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsudot_n_s32))) +svint32_t svsudot_n_s32(svint32_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsudot_s32))) +svint32_t svsudot_s32(svint32_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsudot_lane_s32))) +svint32_t svsudot_lane_s32(svint32_t, svint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svusdot_n_s32))) +svint32_t svusdot_n_s32(svint32_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svusdot_s32))) +svint32_t svusdot_s32(svint32_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svusdot_lane_s32))) +svint32_t svusdot_lane_s32(svint32_t, svuint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svusmmla_s32))) +svint32_t svusmmla_s32(svint32_t, svuint8_t, svint8_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_s32))) svint32_t svmmla(svint32_t, svint8_t, svint8_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_u32))) diff --git a/lib/include/avx512fintrin.h b/lib/include/avx512fintrin.h index f226382cbb..010bcadab0 100644 --- a/lib/include/avx512fintrin.h +++ b/lib/include/avx512fintrin.h @@ -9300,8 +9300,11 @@ _mm512_mask_abs_pd(__m512d __W, __mmask8 __K, __m512d __A) * computations. In vector-reduction arithmetic, the evaluation order is * independent of the order of the input elements of V. - * For floating point types, we always assume the elements are reassociable even - * if -fast-math is off. + * For floating-point intrinsics: + * 1. When using fadd/fmul intrinsics, the order of operations within the + * vector is unspecified (associative math). + * 2. When using fmin/fmax intrinsics, NaN or -0.0 elements within the vector + * produce unspecified results. * Used bisection method. At each step, we partition the vector with previous * step in half, and the operation is performed on its two halves. @@ -9524,75 +9527,49 @@ _mm512_mask_reduce_min_epu32(__mmask16 __M, __m512i __V) { return __builtin_ia32_reduce_umin_d512((__v16si)__V); } -#define _mm512_mask_reduce_operator(op) \ - __m256d __t1 = _mm512_extractf64x4_pd(__V, 0); \ - __m256d __t2 = _mm512_extractf64x4_pd(__V, 1); \ - __m256d __t3 = _mm256_##op(__t1, __t2); \ - __m128d __t4 = _mm256_extractf128_pd(__t3, 0); \ - __m128d __t5 = _mm256_extractf128_pd(__t3, 1); \ - __m128d __t6 = _mm_##op(__t4, __t5); \ - __m128d __t7 = __builtin_shufflevector(__t6, __t6, 1, 0); \ - __m128d __t8 = _mm_##op(__t6, __t7); \ - return __t8[0] - static __inline__ double __DEFAULT_FN_ATTRS512 _mm512_reduce_max_pd(__m512d __V) { - _mm512_mask_reduce_operator(max_pd); + return __builtin_ia32_reduce_fmax_pd512(__V); } static __inline__ double __DEFAULT_FN_ATTRS512 _mm512_reduce_min_pd(__m512d __V) { - _mm512_mask_reduce_operator(min_pd); + return __builtin_ia32_reduce_fmin_pd512(__V); } static __inline__ double __DEFAULT_FN_ATTRS512 _mm512_mask_reduce_max_pd(__mmask8 __M, __m512d __V) { __V = _mm512_mask_mov_pd(_mm512_set1_pd(-__builtin_inf()), __M, __V); - _mm512_mask_reduce_operator(max_pd); + return __builtin_ia32_reduce_fmax_pd512(__V); } static __inline__ double __DEFAULT_FN_ATTRS512 _mm512_mask_reduce_min_pd(__mmask8 __M, __m512d __V) { __V = _mm512_mask_mov_pd(_mm512_set1_pd(__builtin_inf()), __M, __V); - _mm512_mask_reduce_operator(min_pd); + return __builtin_ia32_reduce_fmin_pd512(__V); } -#undef _mm512_mask_reduce_operator - -#define _mm512_mask_reduce_operator(op) \ - __m256 __t1 = (__m256)_mm512_extractf64x4_pd((__m512d)__V, 0); \ - __m256 __t2 = (__m256)_mm512_extractf64x4_pd((__m512d)__V, 1); \ - __m256 __t3 = _mm256_##op(__t1, __t2); \ - __m128 __t4 = _mm256_extractf128_ps(__t3, 0); \ - __m128 __t5 = _mm256_extractf128_ps(__t3, 1); \ - __m128 __t6 = _mm_##op(__t4, __t5); \ - __m128 __t7 = __builtin_shufflevector(__t6, __t6, 2, 3, 0, 1); \ - __m128 __t8 = _mm_##op(__t6, __t7); \ - __m128 __t9 = __builtin_shufflevector(__t8, __t8, 1, 0, 3, 2); \ - __m128 __t10 = _mm_##op(__t8, __t9); \ - return __t10[0] static __inline__ float __DEFAULT_FN_ATTRS512 _mm512_reduce_max_ps(__m512 __V) { - _mm512_mask_reduce_operator(max_ps); + return __builtin_ia32_reduce_fmax_ps512(__V); } static __inline__ float __DEFAULT_FN_ATTRS512 _mm512_reduce_min_ps(__m512 __V) { - _mm512_mask_reduce_operator(min_ps); + return __builtin_ia32_reduce_fmin_ps512(__V); } static __inline__ float __DEFAULT_FN_ATTRS512 _mm512_mask_reduce_max_ps(__mmask16 __M, __m512 __V) { __V = _mm512_mask_mov_ps(_mm512_set1_ps(-__builtin_inff()), __M, __V); - _mm512_mask_reduce_operator(max_ps); + return __builtin_ia32_reduce_fmax_ps512(__V); } static __inline__ float __DEFAULT_FN_ATTRS512 _mm512_mask_reduce_min_ps(__mmask16 __M, __m512 __V) { __V = _mm512_mask_mov_ps(_mm512_set1_ps(__builtin_inff()), __M, __V); - _mm512_mask_reduce_operator(min_ps); + return __builtin_ia32_reduce_fmin_ps512(__V); } -#undef _mm512_mask_reduce_operator /// Moves the least significant 32 bits of a vector of [16 x i32] to a /// 32-bit signed integer value. @@ -9611,6 +9588,169 @@ _mm512_cvtsi512_si32(__m512i __A) { return __b[0]; } +/// Loads 8 double-precision (64-bit) floating-point elements stored at memory +/// locations starting at location \a base_addr at packed 32-bit integer indices +/// stored in the lower half of \a vindex scaled by \a scale them in dst. +/// +/// This intrinsic corresponds to the VGATHERDPD instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// dst[i+63:i] := MEM[addr+63:addr] +/// ENDFOR +/// dst[MAX:512] := 0 +/// \endoperation +#define _mm512_i32logather_pd(vindex, base_addr, scale) \ + _mm512_i32gather_pd(_mm512_castsi512_si256(vindex), (base_addr), (scale)) + +/// Loads 8 double-precision (64-bit) floating-point elements from memory +/// starting at location \a base_addr at packed 32-bit integer indices stored in +/// the lower half of \a vindex scaled by \a scale into dst using writemask +/// \a mask (elements are copied from \a src when the corresponding mask bit is +/// not set). +/// +/// This intrinsic corresponds to the VGATHERDPD instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// IF mask[j] +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// dst[i+63:i] := MEM[addr+63:addr] +/// ELSE +/// dst[i+63:i] := src[i+63:i] +/// FI +/// ENDFOR +/// dst[MAX:512] := 0 +/// \endoperation +#define _mm512_mask_i32logather_pd(src, mask, vindex, base_addr, scale) \ + _mm512_mask_i32gather_pd((src), (mask), _mm512_castsi512_si256(vindex), \ + (base_addr), (scale)) + +/// Loads 8 64-bit integer elements from memory starting at location \a base_addr +/// at packed 32-bit integer indices stored in the lower half of \a vindex +/// scaled by \a scale and stores them in dst. +/// +/// This intrinsic corresponds to the VPGATHERDQ instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// dst[i+63:i] := MEM[addr+63:addr] +/// ENDFOR +/// dst[MAX:512] := 0 +/// \endoperation +#define _mm512_i32logather_epi64(vindex, base_addr, scale) \ + _mm512_i32gather_epi64(_mm512_castsi512_si256(vindex), (base_addr), (scale)) + +/// Loads 8 64-bit integer elements from memory starting at location \a base_addr +/// at packed 32-bit integer indices stored in the lower half of \a vindex +/// scaled by \a scale and stores them in dst using writemask \a mask (elements +/// are copied from \a src when the corresponding mask bit is not set). +/// +/// This intrinsic corresponds to the VPGATHERDQ instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// IF mask[j] +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// dst[i+63:i] := MEM[addr+63:addr] +/// ELSE +/// dst[i+63:i] := src[i+63:i] +/// FI +/// ENDFOR +/// dst[MAX:512] := 0 +/// \endoperation +#define _mm512_mask_i32logather_epi64(src, mask, vindex, base_addr, scale) \ + _mm512_mask_i32gather_epi64((src), (mask), _mm512_castsi512_si256(vindex), \ + (base_addr), (scale)) + +/// Stores 8 packed double-precision (64-bit) floating-point elements in \a v1 +/// and to memory locations starting at location \a base_addr at packed 32-bit +/// integer indices stored in \a vindex scaled by \a scale. +/// +/// This intrinsic corresponds to the VSCATTERDPD instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// MEM[addr+63:addr] := v1[i+63:i] +/// ENDFOR +/// \endoperation +#define _mm512_i32loscatter_pd(base_addr, vindex, v1, scale) \ + _mm512_i32scatter_pd((base_addr), _mm512_castsi512_si256(vindex), (v1), (scale)) + +/// Stores 8 packed double-precision (64-bit) floating-point elements in \a v1 +/// to memory locations starting at location \a base_addr at packed 32-bit +/// integer indices stored in \a vindex scaled by \a scale. Only those elements +/// whose corresponding mask bit is set in writemask \a mask are written to +/// memory. +/// +/// This intrinsic corresponds to the VSCATTERDPD instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// IF mask[j] +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// MEM[addr+63:addr] := a[i+63:i] +/// FI +/// ENDFOR +/// \endoperation +#define _mm512_mask_i32loscatter_pd(base_addr, mask, vindex, v1, scale) \ + _mm512_mask_i32scatter_pd((base_addr), (mask), \ + _mm512_castsi512_si256(vindex), (v1), (scale)) + +/// Stores 8 packed 64-bit integer elements located in \a v1 and stores them in +/// memory locations starting at location \a base_addr at packed 32-bit integer +/// indices stored in \a vindex scaled by \a scale. +/// +/// This intrinsic corresponds to the VPSCATTERDQ instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// MEM[addr+63:addr] := a[i+63:i] +/// ENDFOR +/// \endoperation +#define _mm512_i32loscatter_epi64(base_addr, vindex, v1, scale) \ + _mm512_i32scatter_epi64((base_addr), \ + _mm512_castsi512_si256(vindex), (v1), (scale)) + +/// Stores 8 packed 64-bit integer elements located in a and stores them in +/// memory locations starting at location \a base_addr at packed 32-bit integer +/// indices stored in \a vindex scaled by scale using writemask \a mask (elements +/// whose corresponding mask bit is not set are not written to memory). +/// +/// This intrinsic corresponds to the VPSCATTERDQ instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// IF mask[j] +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// MEM[addr+63:addr] := a[i+63:i] +/// FI +/// ENDFOR +/// \endoperation +#define _mm512_mask_i32loscatter_epi64(base_addr, mask, vindex, v1, scale) \ + _mm512_mask_i32scatter_epi64((base_addr), (mask), \ + _mm512_castsi512_si256(vindex), (v1), (scale)) + #undef __DEFAULT_FN_ATTRS512 #undef __DEFAULT_FN_ATTRS128 #undef __DEFAULT_FN_ATTRS diff --git a/lib/include/builtins.h b/lib/include/builtins.h new file mode 100644 index 0000000000..65095861ca --- /dev/null +++ b/lib/include/builtins.h @@ -0,0 +1,16 @@ +/*===---- builtins.h - Standard header for extra builtins -----------------===*\ + * + * Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. + * See https://llvm.org/LICENSE.txt for license information. + * SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception + * +\*===----------------------------------------------------------------------===*/ + +/// Some legacy compilers have builtin definitions in a file named builtins.h. +/// This header file has been added to allow compatibility with code that was +/// written for those compilers. Code may have an include line for this file +/// and to avoid an error an empty file with this name is provided. +#ifndef __BUILTINS_H +#define __BUILTINS_H + +#endif /* __BUILTINS_H */ diff --git a/lib/include/cuda_wrappers/complex b/lib/include/cuda_wrappers/complex index 11d40a82a8..e6805b6044 100644 --- a/lib/include/cuda_wrappers/complex +++ b/lib/include/cuda_wrappers/complex @@ -72,8 +72,16 @@ #define _GLIBCXX_USE_C99_COMPLEX 0 #define _GLIBCXX_USE_C99_COMPLEX_TR1 0 +// Work around a compatibility issue with libstdc++ 11.1.0 +// https://bugs.llvm.org/show_bug.cgi?id=50383 +#pragma push_macro("__failed_assertion") +#if _GLIBCXX_RELEASE == 11 +#define __failed_assertion __cuda_failed_assertion +#endif + #include_next +#pragma pop_macro("__failed_assertion") #pragma pop_macro("_GLIBCXX_USE_C99_COMPLEX_TR1") #pragma pop_macro("_GLIBCXX_USE_C99_COMPLEX") diff --git a/lib/include/hexagon_circ_brev_intrinsics.h b/lib/include/hexagon_circ_brev_intrinsics.h new file mode 100644 index 0000000000..c53786d3c3 --- /dev/null +++ b/lib/include/hexagon_circ_brev_intrinsics.h @@ -0,0 +1,298 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _HEXAGON_CIRC_BREV_INTRINSICS_H_ +#define _HEXAGON_CIRC_BREV_INTRINSICS_H_ 1 + +#include +#include + +/* Circular Load */ +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_load_update_D(Word64 dst, Word64 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_load_update_D(dest,ptr,incr,bufsize,K) \ + { ptr = (int64_t *) HEXAGON_circ_ldd (ptr, &(dest), ((((K)+1)<<24)|((bufsize)<<3)), ((incr)*8)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_load_update_W(Word32 dst, Word32 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_load_update_W(dest,ptr,incr,bufsize,K) \ + { ptr = (int *) HEXAGON_circ_ldw (ptr, &(dest), (((K)<<24)|((bufsize)<<2)), ((incr)*4)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_load_update_H(Word16 dst, Word16 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_load_update_H(dest,ptr,incr,bufsize,K) \ + { ptr = (int16_t *) HEXAGON_circ_ldh (ptr, &(dest), ((((K)-1)<<24)|((bufsize)<<1)), ((incr)*2)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_load_update_UH( UWord16 dst, UWord16 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_load_update_UH(dest,ptr,incr,bufsize,K) \ + { ptr = (uint16_t *) HEXAGON_circ_lduh (ptr, &(dest), ((((K)-1)<<24)|((bufsize)<<1)), ((incr)*2)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_load_update_B(Word8 dst, Word8 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_load_update_B(dest,ptr,incr,bufsize,K) \ + { ptr = (int8_t *) HEXAGON_circ_ldb (ptr, &(dest), ((((K)-2)<<24)|(bufsize)), incr); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_load_update_UB(UWord8 dst, UWord8 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_load_update_UB(dest,ptr,incr,bufsize,K) \ + { ptr = (uint8_t *) HEXAGON_circ_ldub (ptr, &(dest), ((((K)-2)<<24)|(bufsize)), incr); } + +/* Circular Store */ +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_store_update_D(Word64 *src, Word64 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_store_update_D(src,ptr,incr,bufsize,K) \ + { ptr = (int64_t *) HEXAGON_circ_std (ptr, src, ((((K)+1)<<24)|((bufsize)<<3)), ((incr)*8)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_store_update_W(Word32 *src, Word32 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_store_update_W(src,ptr,incr,bufsize,K) \ + { ptr = (int *) HEXAGON_circ_stw (ptr, src, (((K)<<24)|((bufsize)<<2)), ((incr)*4)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_store_update_HL(Word16 *src, Word16 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_store_update_HL(src,ptr,incr,bufsize,K) \ + { ptr = (int16_t *) HEXAGON_circ_sth (ptr, src, ((((K)-1)<<24)|((bufsize)<<1)), ((incr)*2)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_store_update_HH(Word16 *src, Word16 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_store_update_HH(src,ptr,incr,bufsize,K) \ + { ptr = (int16_t *) HEXAGON_circ_sthhi (ptr, src, ((((K)-1)<<24)|((bufsize)<<1)), ((incr)*2)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_store_update_B(Word8 *src, Word8 *ptr, UWord32 I4, UWord32 bufsize, UWord64 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_store_update_B(src,ptr,incr,bufsize,K) \ + { ptr = (int8_t *) HEXAGON_circ_stb (ptr, src, ((((K)-2)<<24)|(bufsize)), incr); } + + +/* Bit Reverse Load */ +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_load_update_D(Word64 dst, Word64 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_load_update_D(dest,ptr,log2bufsize) \ + { ptr = (int64_t *) HEXAGON_brev_ldd (ptr, &(dest), (1<<(16-((log2bufsize) + 3)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_load_update_W(Word32 dst, Word32 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_load_update_W(dest,ptr,log2bufsize) \ + { ptr = (int *) HEXAGON_brev_ldw (ptr, &(dest), (1<<(16-((log2bufsize) + 2)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_load_update_H(Word16 dst, Word16 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_load_update_H(dest,ptr,log2bufsize) \ + { ptr = (int16_t *) HEXAGON_brev_ldh (ptr, &(dest), (1<<(16-((log2bufsize) + 1)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_load_update_UH(UWord16 dst, UWord16 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_load_update_UH(dest,ptr,log2bufsize) \ + { ptr = (uint16_t *) HEXAGON_brev_lduh (ptr, &(dest), (1<<(16-((log2bufsize) + 1)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_load_update_B(Word8 dst, Word8 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_load_update_B(dest,ptr,log2bufsize) \ + { ptr = (int8_t *) HEXAGON_brev_ldb (ptr, &(dest), (1<<(16-((log2bufsize))))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_load_update_UB(UWord8 dst, UWord8 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_load_update_UB(dest,ptr,log2bufsize) \ + { ptr = (uint8_t *) HEXAGON_brev_ldub (ptr, &(dest), (1<<(16-((log2bufsize))))); } + +/* Bit Reverse Store */ + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_store_update_D(Word64 *src, Word64 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_store_update_D(src,ptr,log2bufsize) \ + { ptr = (int64_t *) HEXAGON_brev_std (ptr, src, (1<<(16-((log2bufsize) + 3)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_store_update_W(Word32 *src, Word32 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_store_update_W(src,ptr,log2bufsize) \ + { ptr = (int *) HEXAGON_brev_stw (ptr, src, (1<<(16-((log2bufsize) + 2)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_store_update_HL(Word16 *src, Word16 *ptr, Word32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_store_update_HL(src,ptr,log2bufsize) \ + { ptr = (int16_t *) HEXAGON_brev_sth (ptr, src, (1<<(16-((log2bufsize) + 1)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_store_update_HH(Word16 *src, Word16 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_store_update_HH(src,ptr,log2bufsize) \ + { ptr = (int16_t *) HEXAGON_brev_sthhi (ptr, src, (1<<(16-((log2bufsize) + 1)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_store_update_B(Word8 *src, Word8 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_store_update_B(src,ptr,log2bufsize) \ + { ptr = (int8_t *) HEXAGON_brev_stb (ptr, src, (1<<(16-((log2bufsize))))); } + + +#define HEXAGON_circ_ldd __builtin_circ_ldd +#define HEXAGON_circ_ldw __builtin_circ_ldw +#define HEXAGON_circ_ldh __builtin_circ_ldh +#define HEXAGON_circ_lduh __builtin_circ_lduh +#define HEXAGON_circ_ldb __builtin_circ_ldb +#define HEXAGON_circ_ldub __builtin_circ_ldub + + +#define HEXAGON_circ_std __builtin_circ_std +#define HEXAGON_circ_stw __builtin_circ_stw +#define HEXAGON_circ_sth __builtin_circ_sth +#define HEXAGON_circ_sthhi __builtin_circ_sthhi +#define HEXAGON_circ_stb __builtin_circ_stb + + +#define HEXAGON_brev_ldd __builtin_brev_ldd +#define HEXAGON_brev_ldw __builtin_brev_ldw +#define HEXAGON_brev_ldh __builtin_brev_ldh +#define HEXAGON_brev_lduh __builtin_brev_lduh +#define HEXAGON_brev_ldb __builtin_brev_ldb +#define HEXAGON_brev_ldub __builtin_brev_ldub + +#define HEXAGON_brev_std __builtin_brev_std +#define HEXAGON_brev_stw __builtin_brev_stw +#define HEXAGON_brev_sth __builtin_brev_sth +#define HEXAGON_brev_sthhi __builtin_brev_sthhi +#define HEXAGON_brev_stb __builtin_brev_stb + +#ifdef __HVX__ +/* ========================================================================== + Assembly Syntax: if (Qt) vmem(Rt+#0) = Vs + C Intrinsic Prototype: void Q6_vmaskedstoreq_QAV(HVX_VectorPred Qt, HVX_VectorAddress A, HVX_Vector Vs) + Instruction Type: COPROC_VMEM + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmaskedstoreq_QAV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaskedstoreq) + +/* ========================================================================== + Assembly Syntax: if (!Qt) vmem(Rt+#0) = Vs + C Intrinsic Prototype: void Q6_vmaskedstorenq_QAV(HVX_VectorPred Qt, HVX_VectorAddress A, HVX_Vector Vs) + Instruction Type: COPROC_VMEM + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmaskedstorenq_QAV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaskedstorenq) + +/* ========================================================================== + Assembly Syntax: if (Qt) vmem(Rt+#0):nt = Vs + C Intrinsic Prototype: void Q6_vmaskedstorentq_QAV(HVX_VectorPred Qt, HVX_VectorAddress A, HVX_Vector Vs) + Instruction Type: COPROC_VMEM + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmaskedstorentq_QAV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaskedstorentq) + +/* ========================================================================== + Assembly Syntax: if (!Qt) vmem(Rt+#0):nt = Vs + C Intrinsic Prototype: void Q6_vmaskedstorentnq_QAV(HVX_VectorPred Qt, HVX_VectorAddress A, HVX_Vector Vs) + Instruction Type: COPROC_VMEM + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmaskedstorentnq_QAV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaskedstorentnq) + +#endif + + +#endif /* #ifndef _HEXAGON_CIRC_BREV_INTRINSICS_H_ */ + +#ifdef __NOT_DEFINED__ +/*** comment block template ***/ +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: ReturnType Intrinsic(ParamType Rs, ParamType Rt) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#endif /*** __NOT_DEFINED__ ***/ diff --git a/lib/include/hexagon_protos.h b/lib/include/hexagon_protos.h new file mode 100644 index 0000000000..cdffd93bb8 --- /dev/null +++ b/lib/include/hexagon_protos.h @@ -0,0 +1,8450 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +// Automatically generated file, do not edit! +//===----------------------------------------------------------------------===// + + + +#ifndef __HEXAGON_PROTOS_H_ +#define __HEXAGON_PROTOS_H_ 1 + +/* ========================================================================== + Assembly Syntax: Rd32=abs(Rs32) + C Intrinsic Prototype: Word32 Q6_R_abs_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_abs_R __builtin_HEXAGON_A2_abs + +/* ========================================================================== + Assembly Syntax: Rdd32=abs(Rss32) + C Intrinsic Prototype: Word64 Q6_P_abs_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_abs_P __builtin_HEXAGON_A2_absp + +/* ========================================================================== + Assembly Syntax: Rd32=abs(Rs32):sat + C Intrinsic Prototype: Word32 Q6_R_abs_R_sat(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_abs_R_sat __builtin_HEXAGON_A2_abssat + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_add_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_add_RR __builtin_HEXAGON_A2_add + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.h,Rs32.h):<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RhRh_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RhRh_s16 __builtin_HEXAGON_A2_addh_h16_hh + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.h,Rs32.l):<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RhRl_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RhRl_s16 __builtin_HEXAGON_A2_addh_h16_hl + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.h):<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RlRh_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRh_s16 __builtin_HEXAGON_A2_addh_h16_lh + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.l):<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RlRl_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRl_s16 __builtin_HEXAGON_A2_addh_h16_ll + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.h,Rs32.h):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RhRh_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RhRh_sat_s16 __builtin_HEXAGON_A2_addh_h16_sat_hh + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.h,Rs32.l):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RhRl_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RhRl_sat_s16 __builtin_HEXAGON_A2_addh_h16_sat_hl + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.h):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RlRh_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRh_sat_s16 __builtin_HEXAGON_A2_addh_h16_sat_lh + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.l):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RlRl_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRl_sat_s16 __builtin_HEXAGON_A2_addh_h16_sat_ll + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.h) + C Intrinsic Prototype: Word32 Q6_R_add_RlRh(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRh __builtin_HEXAGON_A2_addh_l16_hl + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.l) + C Intrinsic Prototype: Word32 Q6_R_add_RlRl(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRl __builtin_HEXAGON_A2_addh_l16_ll + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.h):sat + C Intrinsic Prototype: Word32 Q6_R_add_RlRh_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRh_sat __builtin_HEXAGON_A2_addh_l16_sat_hl + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.l):sat + C Intrinsic Prototype: Word32 Q6_R_add_RlRl_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRl_sat __builtin_HEXAGON_A2_addh_l16_sat_ll + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rs32,#s16) + C Intrinsic Prototype: Word32 Q6_R_add_RI(Word32 Rs, Word32 Is16) + Instruction Type: ALU32_ADDI + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_add_RI __builtin_HEXAGON_A2_addi + +/* ========================================================================== + Assembly Syntax: Rdd32=add(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_add_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_add_PP __builtin_HEXAGON_A2_addp + +/* ========================================================================== + Assembly Syntax: Rdd32=add(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_add_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_add_PP_sat __builtin_HEXAGON_A2_addpsat + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rs32,Rt32):sat + C Intrinsic Prototype: Word32 Q6_R_add_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_add_RR_sat __builtin_HEXAGON_A2_addsat + +/* ========================================================================== + Assembly Syntax: Rdd32=add(Rs32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_add_RP(Word32 Rs, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_add_RP __builtin_HEXAGON_A2_addsp + +/* ========================================================================== + Assembly Syntax: Rd32=and(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_and_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_and_RR __builtin_HEXAGON_A2_and + +/* ========================================================================== + Assembly Syntax: Rd32=and(Rs32,#s10) + C Intrinsic Prototype: Word32 Q6_R_and_RI(Word32 Rs, Word32 Is10) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_and_RI __builtin_HEXAGON_A2_andir + +/* ========================================================================== + Assembly Syntax: Rdd32=and(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_and_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_and_PP __builtin_HEXAGON_A2_andp + +/* ========================================================================== + Assembly Syntax: Rd32=aslh(Rs32) + C Intrinsic Prototype: Word32 Q6_R_aslh_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_aslh_R __builtin_HEXAGON_A2_aslh + +/* ========================================================================== + Assembly Syntax: Rd32=asrh(Rs32) + C Intrinsic Prototype: Word32 Q6_R_asrh_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_asrh_R __builtin_HEXAGON_A2_asrh + +/* ========================================================================== + Assembly Syntax: Rd32=combine(Rt32.h,Rs32.h) + C Intrinsic Prototype: Word32 Q6_R_combine_RhRh(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_combine_RhRh __builtin_HEXAGON_A2_combine_hh + +/* ========================================================================== + Assembly Syntax: Rd32=combine(Rt32.h,Rs32.l) + C Intrinsic Prototype: Word32 Q6_R_combine_RhRl(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_combine_RhRl __builtin_HEXAGON_A2_combine_hl + +/* ========================================================================== + Assembly Syntax: Rd32=combine(Rt32.l,Rs32.h) + C Intrinsic Prototype: Word32 Q6_R_combine_RlRh(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_combine_RlRh __builtin_HEXAGON_A2_combine_lh + +/* ========================================================================== + Assembly Syntax: Rd32=combine(Rt32.l,Rs32.l) + C Intrinsic Prototype: Word32 Q6_R_combine_RlRl(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_combine_RlRl __builtin_HEXAGON_A2_combine_ll + +/* ========================================================================== + Assembly Syntax: Rdd32=combine(#s8,#S8) + C Intrinsic Prototype: Word64 Q6_P_combine_II(Word32 Is8, Word32 IS8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_combine_II __builtin_HEXAGON_A2_combineii + +/* ========================================================================== + Assembly Syntax: Rdd32=combine(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_combine_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_combine_RR __builtin_HEXAGON_A2_combinew + +/* ========================================================================== + Assembly Syntax: Rd32=max(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_max_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_max_RR __builtin_HEXAGON_A2_max + +/* ========================================================================== + Assembly Syntax: Rdd32=max(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_max_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_max_PP __builtin_HEXAGON_A2_maxp + +/* ========================================================================== + Assembly Syntax: Rd32=maxu(Rs32,Rt32) + C Intrinsic Prototype: UWord32 Q6_R_maxu_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_maxu_RR __builtin_HEXAGON_A2_maxu + +/* ========================================================================== + Assembly Syntax: Rdd32=maxu(Rss32,Rtt32) + C Intrinsic Prototype: UWord64 Q6_P_maxu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_maxu_PP __builtin_HEXAGON_A2_maxup + +/* ========================================================================== + Assembly Syntax: Rd32=min(Rt32,Rs32) + C Intrinsic Prototype: Word32 Q6_R_min_RR(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_min_RR __builtin_HEXAGON_A2_min + +/* ========================================================================== + Assembly Syntax: Rdd32=min(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_min_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_min_PP __builtin_HEXAGON_A2_minp + +/* ========================================================================== + Assembly Syntax: Rd32=minu(Rt32,Rs32) + C Intrinsic Prototype: UWord32 Q6_R_minu_RR(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_minu_RR __builtin_HEXAGON_A2_minu + +/* ========================================================================== + Assembly Syntax: Rdd32=minu(Rtt32,Rss32) + C Intrinsic Prototype: UWord64 Q6_P_minu_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_minu_PP __builtin_HEXAGON_A2_minup + +/* ========================================================================== + Assembly Syntax: Rd32=neg(Rs32) + C Intrinsic Prototype: Word32 Q6_R_neg_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_neg_R __builtin_HEXAGON_A2_neg + +/* ========================================================================== + Assembly Syntax: Rdd32=neg(Rss32) + C Intrinsic Prototype: Word64 Q6_P_neg_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_neg_P __builtin_HEXAGON_A2_negp + +/* ========================================================================== + Assembly Syntax: Rd32=neg(Rs32):sat + C Intrinsic Prototype: Word32 Q6_R_neg_R_sat(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_neg_R_sat __builtin_HEXAGON_A2_negsat + +/* ========================================================================== + Assembly Syntax: Rd32=not(Rs32) + C Intrinsic Prototype: Word32 Q6_R_not_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_not_R __builtin_HEXAGON_A2_not + +/* ========================================================================== + Assembly Syntax: Rdd32=not(Rss32) + C Intrinsic Prototype: Word64 Q6_P_not_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_not_P __builtin_HEXAGON_A2_notp + +/* ========================================================================== + Assembly Syntax: Rd32=or(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_or_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_or_RR __builtin_HEXAGON_A2_or + +/* ========================================================================== + Assembly Syntax: Rd32=or(Rs32,#s10) + C Intrinsic Prototype: Word32 Q6_R_or_RI(Word32 Rs, Word32 Is10) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_or_RI __builtin_HEXAGON_A2_orir + +/* ========================================================================== + Assembly Syntax: Rdd32=or(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_or_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_or_PP __builtin_HEXAGON_A2_orp + +/* ========================================================================== + Assembly Syntax: Rd32=round(Rss32):sat + C Intrinsic Prototype: Word32 Q6_R_round_P_sat(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_round_P_sat __builtin_HEXAGON_A2_roundsat + +/* ========================================================================== + Assembly Syntax: Rd32=sat(Rss32) + C Intrinsic Prototype: Word32 Q6_R_sat_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sat_P __builtin_HEXAGON_A2_sat + +/* ========================================================================== + Assembly Syntax: Rd32=satb(Rs32) + C Intrinsic Prototype: Word32 Q6_R_satb_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_satb_R __builtin_HEXAGON_A2_satb + +/* ========================================================================== + Assembly Syntax: Rd32=sath(Rs32) + C Intrinsic Prototype: Word32 Q6_R_sath_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sath_R __builtin_HEXAGON_A2_sath + +/* ========================================================================== + Assembly Syntax: Rd32=satub(Rs32) + C Intrinsic Prototype: Word32 Q6_R_satub_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_satub_R __builtin_HEXAGON_A2_satub + +/* ========================================================================== + Assembly Syntax: Rd32=satuh(Rs32) + C Intrinsic Prototype: Word32 Q6_R_satuh_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_satuh_R __builtin_HEXAGON_A2_satuh + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32,Rs32) + C Intrinsic Prototype: Word32 Q6_R_sub_RR(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_sub_RR __builtin_HEXAGON_A2_sub + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.h,Rs32.h):<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RhRh_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RhRh_s16 __builtin_HEXAGON_A2_subh_h16_hh + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.h,Rs32.l):<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RhRl_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RhRl_s16 __builtin_HEXAGON_A2_subh_h16_hl + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.h):<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RlRh_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRh_s16 __builtin_HEXAGON_A2_subh_h16_lh + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.l):<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RlRl_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRl_s16 __builtin_HEXAGON_A2_subh_h16_ll + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.h,Rs32.h):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RhRh_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RhRh_sat_s16 __builtin_HEXAGON_A2_subh_h16_sat_hh + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.h,Rs32.l):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RhRl_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RhRl_sat_s16 __builtin_HEXAGON_A2_subh_h16_sat_hl + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.h):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RlRh_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRh_sat_s16 __builtin_HEXAGON_A2_subh_h16_sat_lh + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.l):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RlRl_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRl_sat_s16 __builtin_HEXAGON_A2_subh_h16_sat_ll + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.h) + C Intrinsic Prototype: Word32 Q6_R_sub_RlRh(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRh __builtin_HEXAGON_A2_subh_l16_hl + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.l) + C Intrinsic Prototype: Word32 Q6_R_sub_RlRl(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRl __builtin_HEXAGON_A2_subh_l16_ll + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.h):sat + C Intrinsic Prototype: Word32 Q6_R_sub_RlRh_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRh_sat __builtin_HEXAGON_A2_subh_l16_sat_hl + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.l):sat + C Intrinsic Prototype: Word32 Q6_R_sub_RlRl_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRl_sat __builtin_HEXAGON_A2_subh_l16_sat_ll + +/* ========================================================================== + Assembly Syntax: Rdd32=sub(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_sub_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_sub_PP __builtin_HEXAGON_A2_subp + +/* ========================================================================== + Assembly Syntax: Rd32=sub(#s10,Rs32) + C Intrinsic Prototype: Word32 Q6_R_sub_IR(Word32 Is10, Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_sub_IR __builtin_HEXAGON_A2_subri + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32,Rs32):sat + C Intrinsic Prototype: Word32 Q6_R_sub_RR_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_sub_RR_sat __builtin_HEXAGON_A2_subsat + +/* ========================================================================== + Assembly Syntax: Rd32=vaddh(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_vaddh_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vaddh_RR __builtin_HEXAGON_A2_svaddh + +/* ========================================================================== + Assembly Syntax: Rd32=vaddh(Rs32,Rt32):sat + C Intrinsic Prototype: Word32 Q6_R_vaddh_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vaddh_RR_sat __builtin_HEXAGON_A2_svaddhs + +/* ========================================================================== + Assembly Syntax: Rd32=vadduh(Rs32,Rt32):sat + C Intrinsic Prototype: Word32 Q6_R_vadduh_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vadduh_RR_sat __builtin_HEXAGON_A2_svadduhs + +/* ========================================================================== + Assembly Syntax: Rd32=vavgh(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_vavgh_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vavgh_RR __builtin_HEXAGON_A2_svavgh + +/* ========================================================================== + Assembly Syntax: Rd32=vavgh(Rs32,Rt32):rnd + C Intrinsic Prototype: Word32 Q6_R_vavgh_RR_rnd(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vavgh_RR_rnd __builtin_HEXAGON_A2_svavghs + +/* ========================================================================== + Assembly Syntax: Rd32=vnavgh(Rt32,Rs32) + C Intrinsic Prototype: Word32 Q6_R_vnavgh_RR(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vnavgh_RR __builtin_HEXAGON_A2_svnavgh + +/* ========================================================================== + Assembly Syntax: Rd32=vsubh(Rt32,Rs32) + C Intrinsic Prototype: Word32 Q6_R_vsubh_RR(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vsubh_RR __builtin_HEXAGON_A2_svsubh + +/* ========================================================================== + Assembly Syntax: Rd32=vsubh(Rt32,Rs32):sat + C Intrinsic Prototype: Word32 Q6_R_vsubh_RR_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vsubh_RR_sat __builtin_HEXAGON_A2_svsubhs + +/* ========================================================================== + Assembly Syntax: Rd32=vsubuh(Rt32,Rs32):sat + C Intrinsic Prototype: Word32 Q6_R_vsubuh_RR_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vsubuh_RR_sat __builtin_HEXAGON_A2_svsubuhs + +/* ========================================================================== + Assembly Syntax: Rd32=swiz(Rs32) + C Intrinsic Prototype: Word32 Q6_R_swiz_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_swiz_R __builtin_HEXAGON_A2_swiz + +/* ========================================================================== + Assembly Syntax: Rd32=sxtb(Rs32) + C Intrinsic Prototype: Word32 Q6_R_sxtb_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_sxtb_R __builtin_HEXAGON_A2_sxtb + +/* ========================================================================== + Assembly Syntax: Rd32=sxth(Rs32) + C Intrinsic Prototype: Word32 Q6_R_sxth_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_sxth_R __builtin_HEXAGON_A2_sxth + +/* ========================================================================== + Assembly Syntax: Rdd32=sxtw(Rs32) + C Intrinsic Prototype: Word64 Q6_P_sxtw_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_sxtw_R __builtin_HEXAGON_A2_sxtw + +/* ========================================================================== + Assembly Syntax: Rd32=Rs32 + C Intrinsic Prototype: Word32 Q6_R_equals_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_equals_R __builtin_HEXAGON_A2_tfr + +/* ========================================================================== + Assembly Syntax: Rx32.h=#u16 + C Intrinsic Prototype: Word32 Q6_Rh_equals_I(Word32 Rx, Word32 Iu16) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Rh_equals_I __builtin_HEXAGON_A2_tfrih + +/* ========================================================================== + Assembly Syntax: Rx32.l=#u16 + C Intrinsic Prototype: Word32 Q6_Rl_equals_I(Word32 Rx, Word32 Iu16) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Rl_equals_I __builtin_HEXAGON_A2_tfril + +/* ========================================================================== + Assembly Syntax: Rdd32=Rss32 + C Intrinsic Prototype: Word64 Q6_P_equals_P(Word64 Rss) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_equals_P __builtin_HEXAGON_A2_tfrp + +/* ========================================================================== + Assembly Syntax: Rdd32=#s8 + C Intrinsic Prototype: Word64 Q6_P_equals_I(Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_equals_I __builtin_HEXAGON_A2_tfrpi + +/* ========================================================================== + Assembly Syntax: Rd32=#s16 + C Intrinsic Prototype: Word32 Q6_R_equals_I(Word32 Is16) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_equals_I __builtin_HEXAGON_A2_tfrsi + +/* ========================================================================== + Assembly Syntax: Rdd32=vabsh(Rss32) + C Intrinsic Prototype: Word64 Q6_P_vabsh_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsh_P __builtin_HEXAGON_A2_vabsh + +/* ========================================================================== + Assembly Syntax: Rdd32=vabsh(Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vabsh_P_sat(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsh_P_sat __builtin_HEXAGON_A2_vabshsat + +/* ========================================================================== + Assembly Syntax: Rdd32=vabsw(Rss32) + C Intrinsic Prototype: Word64 Q6_P_vabsw_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsw_P __builtin_HEXAGON_A2_vabsw + +/* ========================================================================== + Assembly Syntax: Rdd32=vabsw(Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vabsw_P_sat(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsw_P_sat __builtin_HEXAGON_A2_vabswsat + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddb(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vaddb_PP(Word64 Rss, Word64 Rtt) + Instruction Type: MAPPING + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_vaddb_PP __builtin_HEXAGON_A2_vaddb_map + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vaddh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaddh_PP __builtin_HEXAGON_A2_vaddh + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vaddh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaddh_PP_sat __builtin_HEXAGON_A2_vaddhs + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddub(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vaddub_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaddub_PP __builtin_HEXAGON_A2_vaddub + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddub(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vaddub_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaddub_PP_sat __builtin_HEXAGON_A2_vaddubs + +/* ========================================================================== + Assembly Syntax: Rdd32=vadduh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vadduh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vadduh_PP_sat __builtin_HEXAGON_A2_vadduhs + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vaddw_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaddw_PP __builtin_HEXAGON_A2_vaddw + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddw(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vaddw_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaddw_PP_sat __builtin_HEXAGON_A2_vaddws + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vavgh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgh_PP __builtin_HEXAGON_A2_vavgh + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgh(Rss32,Rtt32):crnd + C Intrinsic Prototype: Word64 Q6_P_vavgh_PP_crnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgh_PP_crnd __builtin_HEXAGON_A2_vavghcr + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgh(Rss32,Rtt32):rnd + C Intrinsic Prototype: Word64 Q6_P_vavgh_PP_rnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgh_PP_rnd __builtin_HEXAGON_A2_vavghr + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgub(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vavgub_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgub_PP __builtin_HEXAGON_A2_vavgub + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgub(Rss32,Rtt32):rnd + C Intrinsic Prototype: Word64 Q6_P_vavgub_PP_rnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgub_PP_rnd __builtin_HEXAGON_A2_vavgubr + +/* ========================================================================== + Assembly Syntax: Rdd32=vavguh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vavguh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavguh_PP __builtin_HEXAGON_A2_vavguh + +/* ========================================================================== + Assembly Syntax: Rdd32=vavguh(Rss32,Rtt32):rnd + C Intrinsic Prototype: Word64 Q6_P_vavguh_PP_rnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavguh_PP_rnd __builtin_HEXAGON_A2_vavguhr + +/* ========================================================================== + Assembly Syntax: Rdd32=vavguw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vavguw_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavguw_PP __builtin_HEXAGON_A2_vavguw + +/* ========================================================================== + Assembly Syntax: Rdd32=vavguw(Rss32,Rtt32):rnd + C Intrinsic Prototype: Word64 Q6_P_vavguw_PP_rnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavguw_PP_rnd __builtin_HEXAGON_A2_vavguwr + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vavgw_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgw_PP __builtin_HEXAGON_A2_vavgw + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgw(Rss32,Rtt32):crnd + C Intrinsic Prototype: Word64 Q6_P_vavgw_PP_crnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgw_PP_crnd __builtin_HEXAGON_A2_vavgwcr + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgw(Rss32,Rtt32):rnd + C Intrinsic Prototype: Word64 Q6_P_vavgw_PP_rnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgw_PP_rnd __builtin_HEXAGON_A2_vavgwr + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpb.eq(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmpb_eq_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpb_eq_PP __builtin_HEXAGON_A2_vcmpbeq + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpb.gtu(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmpb_gtu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpb_gtu_PP __builtin_HEXAGON_A2_vcmpbgtu + +/* ========================================================================== + Assembly Syntax: Pd4=vcmph.eq(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmph_eq_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmph_eq_PP __builtin_HEXAGON_A2_vcmpheq + +/* ========================================================================== + Assembly Syntax: Pd4=vcmph.gt(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmph_gt_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmph_gt_PP __builtin_HEXAGON_A2_vcmphgt + +/* ========================================================================== + Assembly Syntax: Pd4=vcmph.gtu(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmph_gtu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmph_gtu_PP __builtin_HEXAGON_A2_vcmphgtu + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpw.eq(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmpw_eq_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpw_eq_PP __builtin_HEXAGON_A2_vcmpweq + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpw.gt(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmpw_gt_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpw_gt_PP __builtin_HEXAGON_A2_vcmpwgt + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpw.gtu(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmpw_gtu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpw_gtu_PP __builtin_HEXAGON_A2_vcmpwgtu + +/* ========================================================================== + Assembly Syntax: Rdd32=vconj(Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vconj_P_sat(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vconj_P_sat __builtin_HEXAGON_A2_vconj + +/* ========================================================================== + Assembly Syntax: Rdd32=vmaxb(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vmaxb_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmaxb_PP __builtin_HEXAGON_A2_vmaxb + +/* ========================================================================== + Assembly Syntax: Rdd32=vmaxh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vmaxh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmaxh_PP __builtin_HEXAGON_A2_vmaxh + +/* ========================================================================== + Assembly Syntax: Rdd32=vmaxub(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vmaxub_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmaxub_PP __builtin_HEXAGON_A2_vmaxub + +/* ========================================================================== + Assembly Syntax: Rdd32=vmaxuh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vmaxuh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmaxuh_PP __builtin_HEXAGON_A2_vmaxuh + +/* ========================================================================== + Assembly Syntax: Rdd32=vmaxuw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vmaxuw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmaxuw_PP __builtin_HEXAGON_A2_vmaxuw + +/* ========================================================================== + Assembly Syntax: Rdd32=vmaxw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vmaxw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmaxw_PP __builtin_HEXAGON_A2_vmaxw + +/* ========================================================================== + Assembly Syntax: Rdd32=vminb(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vminb_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vminb_PP __builtin_HEXAGON_A2_vminb + +/* ========================================================================== + Assembly Syntax: Rdd32=vminh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vminh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vminh_PP __builtin_HEXAGON_A2_vminh + +/* ========================================================================== + Assembly Syntax: Rdd32=vminub(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vminub_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vminub_PP __builtin_HEXAGON_A2_vminub + +/* ========================================================================== + Assembly Syntax: Rdd32=vminuh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vminuh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vminuh_PP __builtin_HEXAGON_A2_vminuh + +/* ========================================================================== + Assembly Syntax: Rdd32=vminuw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vminuw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vminuw_PP __builtin_HEXAGON_A2_vminuw + +/* ========================================================================== + Assembly Syntax: Rdd32=vminw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vminw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vminw_PP __builtin_HEXAGON_A2_vminw + +/* ========================================================================== + Assembly Syntax: Rdd32=vnavgh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vnavgh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vnavgh_PP __builtin_HEXAGON_A2_vnavgh + +/* ========================================================================== + Assembly Syntax: Rdd32=vnavgh(Rtt32,Rss32):crnd:sat + C Intrinsic Prototype: Word64 Q6_P_vnavgh_PP_crnd_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vnavgh_PP_crnd_sat __builtin_HEXAGON_A2_vnavghcr + +/* ========================================================================== + Assembly Syntax: Rdd32=vnavgh(Rtt32,Rss32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vnavgh_PP_rnd_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vnavgh_PP_rnd_sat __builtin_HEXAGON_A2_vnavghr + +/* ========================================================================== + Assembly Syntax: Rdd32=vnavgw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vnavgw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vnavgw_PP __builtin_HEXAGON_A2_vnavgw + +/* ========================================================================== + Assembly Syntax: Rdd32=vnavgw(Rtt32,Rss32):crnd:sat + C Intrinsic Prototype: Word64 Q6_P_vnavgw_PP_crnd_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vnavgw_PP_crnd_sat __builtin_HEXAGON_A2_vnavgwcr + +/* ========================================================================== + Assembly Syntax: Rdd32=vnavgw(Rtt32,Rss32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vnavgw_PP_rnd_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vnavgw_PP_rnd_sat __builtin_HEXAGON_A2_vnavgwr + +/* ========================================================================== + Assembly Syntax: Rdd32=vraddub(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vraddub_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vraddub_PP __builtin_HEXAGON_A2_vraddub + +/* ========================================================================== + Assembly Syntax: Rxx32+=vraddub(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vraddubacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vraddubacc_PP __builtin_HEXAGON_A2_vraddub_acc + +/* ========================================================================== + Assembly Syntax: Rdd32=vrsadub(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrsadub_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrsadub_PP __builtin_HEXAGON_A2_vrsadub + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrsadub(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrsadubacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrsadubacc_PP __builtin_HEXAGON_A2_vrsadub_acc + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubb(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vsubb_PP(Word64 Rss, Word64 Rtt) + Instruction Type: MAPPING + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_vsubb_PP __builtin_HEXAGON_A2_vsubb_map + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsubh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubh_PP __builtin_HEXAGON_A2_vsubh + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubh(Rtt32,Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vsubh_PP_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubh_PP_sat __builtin_HEXAGON_A2_vsubhs + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubub(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsubub_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubub_PP __builtin_HEXAGON_A2_vsubub + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubub(Rtt32,Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vsubub_PP_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubub_PP_sat __builtin_HEXAGON_A2_vsububs + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubuh(Rtt32,Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vsubuh_PP_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubuh_PP_sat __builtin_HEXAGON_A2_vsubuhs + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsubw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubw_PP __builtin_HEXAGON_A2_vsubw + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubw(Rtt32,Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vsubw_PP_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubw_PP_sat __builtin_HEXAGON_A2_vsubws + +/* ========================================================================== + Assembly Syntax: Rd32=xor(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_xor_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_xor_RR __builtin_HEXAGON_A2_xor + +/* ========================================================================== + Assembly Syntax: Rdd32=xor(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_xor_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_xor_PP __builtin_HEXAGON_A2_xorp + +/* ========================================================================== + Assembly Syntax: Rd32=zxtb(Rs32) + C Intrinsic Prototype: Word32 Q6_R_zxtb_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_zxtb_R __builtin_HEXAGON_A2_zxtb + +/* ========================================================================== + Assembly Syntax: Rd32=zxth(Rs32) + C Intrinsic Prototype: Word32 Q6_R_zxth_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_zxth_R __builtin_HEXAGON_A2_zxth + +/* ========================================================================== + Assembly Syntax: Rd32=and(Rt32,~Rs32) + C Intrinsic Prototype: Word32 Q6_R_and_RnR(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_and_RnR __builtin_HEXAGON_A4_andn + +/* ========================================================================== + Assembly Syntax: Rdd32=and(Rtt32,~Rss32) + C Intrinsic Prototype: Word64 Q6_P_and_PnP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_and_PnP __builtin_HEXAGON_A4_andnp + +/* ========================================================================== + Assembly Syntax: Rdd32=bitsplit(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_bitsplit_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_bitsplit_RR __builtin_HEXAGON_A4_bitsplit + +/* ========================================================================== + Assembly Syntax: Rdd32=bitsplit(Rs32,#u5) + C Intrinsic Prototype: Word64 Q6_P_bitsplit_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_bitsplit_RI __builtin_HEXAGON_A4_bitspliti + +/* ========================================================================== + Assembly Syntax: Pd4=boundscheck(Rs32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_boundscheck_RP(Word32 Rs, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_boundscheck_RP __builtin_HEXAGON_A4_boundscheck + +/* ========================================================================== + Assembly Syntax: Pd4=cmpb.eq(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmpb_eq_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmpb_eq_RR __builtin_HEXAGON_A4_cmpbeq + +/* ========================================================================== + Assembly Syntax: Pd4=cmpb.eq(Rs32,#u8) + C Intrinsic Prototype: Byte Q6_p_cmpb_eq_RI(Word32 Rs, Word32 Iu8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmpb_eq_RI __builtin_HEXAGON_A4_cmpbeqi + +/* ========================================================================== + Assembly Syntax: Pd4=cmpb.gt(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmpb_gt_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmpb_gt_RR __builtin_HEXAGON_A4_cmpbgt + +/* ========================================================================== + Assembly Syntax: Pd4=cmpb.gt(Rs32,#s8) + C Intrinsic Prototype: Byte Q6_p_cmpb_gt_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmpb_gt_RI __builtin_HEXAGON_A4_cmpbgti + +/* ========================================================================== + Assembly Syntax: Pd4=cmpb.gtu(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmpb_gtu_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmpb_gtu_RR __builtin_HEXAGON_A4_cmpbgtu + +/* ========================================================================== + Assembly Syntax: Pd4=cmpb.gtu(Rs32,#u7) + C Intrinsic Prototype: Byte Q6_p_cmpb_gtu_RI(Word32 Rs, Word32 Iu7) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmpb_gtu_RI __builtin_HEXAGON_A4_cmpbgtui + +/* ========================================================================== + Assembly Syntax: Pd4=cmph.eq(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmph_eq_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmph_eq_RR __builtin_HEXAGON_A4_cmpheq + +/* ========================================================================== + Assembly Syntax: Pd4=cmph.eq(Rs32,#s8) + C Intrinsic Prototype: Byte Q6_p_cmph_eq_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmph_eq_RI __builtin_HEXAGON_A4_cmpheqi + +/* ========================================================================== + Assembly Syntax: Pd4=cmph.gt(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmph_gt_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmph_gt_RR __builtin_HEXAGON_A4_cmphgt + +/* ========================================================================== + Assembly Syntax: Pd4=cmph.gt(Rs32,#s8) + C Intrinsic Prototype: Byte Q6_p_cmph_gt_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmph_gt_RI __builtin_HEXAGON_A4_cmphgti + +/* ========================================================================== + Assembly Syntax: Pd4=cmph.gtu(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmph_gtu_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmph_gtu_RR __builtin_HEXAGON_A4_cmphgtu + +/* ========================================================================== + Assembly Syntax: Pd4=cmph.gtu(Rs32,#u7) + C Intrinsic Prototype: Byte Q6_p_cmph_gtu_RI(Word32 Rs, Word32 Iu7) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmph_gtu_RI __builtin_HEXAGON_A4_cmphgtui + +/* ========================================================================== + Assembly Syntax: Rdd32=combine(#s8,Rs32) + C Intrinsic Prototype: Word64 Q6_P_combine_IR(Word32 Is8, Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_combine_IR __builtin_HEXAGON_A4_combineir + +/* ========================================================================== + Assembly Syntax: Rdd32=combine(Rs32,#s8) + C Intrinsic Prototype: Word64 Q6_P_combine_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_combine_RI __builtin_HEXAGON_A4_combineri + +/* ========================================================================== + Assembly Syntax: Rd32=cround(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_cround_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cround_RI __builtin_HEXAGON_A4_cround_ri + +/* ========================================================================== + Assembly Syntax: Rd32=cround(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_cround_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cround_RR __builtin_HEXAGON_A4_cround_rr + +/* ========================================================================== + Assembly Syntax: Rd32=modwrap(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_modwrap_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_modwrap_RR __builtin_HEXAGON_A4_modwrapu + +/* ========================================================================== + Assembly Syntax: Rd32=or(Rt32,~Rs32) + C Intrinsic Prototype: Word32 Q6_R_or_RnR(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_or_RnR __builtin_HEXAGON_A4_orn + +/* ========================================================================== + Assembly Syntax: Rdd32=or(Rtt32,~Rss32) + C Intrinsic Prototype: Word64 Q6_P_or_PnP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_or_PnP __builtin_HEXAGON_A4_ornp + +/* ========================================================================== + Assembly Syntax: Rd32=cmp.eq(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_cmp_eq_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_cmp_eq_RR __builtin_HEXAGON_A4_rcmpeq + +/* ========================================================================== + Assembly Syntax: Rd32=cmp.eq(Rs32,#s8) + C Intrinsic Prototype: Word32 Q6_R_cmp_eq_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_cmp_eq_RI __builtin_HEXAGON_A4_rcmpeqi + +/* ========================================================================== + Assembly Syntax: Rd32=!cmp.eq(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_not_cmp_eq_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_not_cmp_eq_RR __builtin_HEXAGON_A4_rcmpneq + +/* ========================================================================== + Assembly Syntax: Rd32=!cmp.eq(Rs32,#s8) + C Intrinsic Prototype: Word32 Q6_R_not_cmp_eq_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_not_cmp_eq_RI __builtin_HEXAGON_A4_rcmpneqi + +/* ========================================================================== + Assembly Syntax: Rd32=round(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_round_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_round_RI __builtin_HEXAGON_A4_round_ri + +/* ========================================================================== + Assembly Syntax: Rd32=round(Rs32,#u5):sat + C Intrinsic Prototype: Word32 Q6_R_round_RI_sat(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_round_RI_sat __builtin_HEXAGON_A4_round_ri_sat + +/* ========================================================================== + Assembly Syntax: Rd32=round(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_round_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_round_RR __builtin_HEXAGON_A4_round_rr + +/* ========================================================================== + Assembly Syntax: Rd32=round(Rs32,Rt32):sat + C Intrinsic Prototype: Word32 Q6_R_round_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_round_RR_sat __builtin_HEXAGON_A4_round_rr_sat + +/* ========================================================================== + Assembly Syntax: Pd4=tlbmatch(Rss32,Rt32) + C Intrinsic Prototype: Byte Q6_p_tlbmatch_PR(Word64 Rss, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_tlbmatch_PR __builtin_HEXAGON_A4_tlbmatch + +/* ========================================================================== + Assembly Syntax: Pd4=any8(vcmpb.eq(Rss32,Rtt32)) + C Intrinsic Prototype: Byte Q6_p_any8_vcmpb_eq_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_any8_vcmpb_eq_PP __builtin_HEXAGON_A4_vcmpbeq_any + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpb.eq(Rss32,#u8) + C Intrinsic Prototype: Byte Q6_p_vcmpb_eq_PI(Word64 Rss, Word32 Iu8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpb_eq_PI __builtin_HEXAGON_A4_vcmpbeqi + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpb.gt(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmpb_gt_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpb_gt_PP __builtin_HEXAGON_A4_vcmpbgt + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpb.gt(Rss32,#s8) + C Intrinsic Prototype: Byte Q6_p_vcmpb_gt_PI(Word64 Rss, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpb_gt_PI __builtin_HEXAGON_A4_vcmpbgti + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpb.gtu(Rss32,#u7) + C Intrinsic Prototype: Byte Q6_p_vcmpb_gtu_PI(Word64 Rss, Word32 Iu7) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpb_gtu_PI __builtin_HEXAGON_A4_vcmpbgtui + +/* ========================================================================== + Assembly Syntax: Pd4=vcmph.eq(Rss32,#s8) + C Intrinsic Prototype: Byte Q6_p_vcmph_eq_PI(Word64 Rss, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmph_eq_PI __builtin_HEXAGON_A4_vcmpheqi + +/* ========================================================================== + Assembly Syntax: Pd4=vcmph.gt(Rss32,#s8) + C Intrinsic Prototype: Byte Q6_p_vcmph_gt_PI(Word64 Rss, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmph_gt_PI __builtin_HEXAGON_A4_vcmphgti + +/* ========================================================================== + Assembly Syntax: Pd4=vcmph.gtu(Rss32,#u7) + C Intrinsic Prototype: Byte Q6_p_vcmph_gtu_PI(Word64 Rss, Word32 Iu7) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmph_gtu_PI __builtin_HEXAGON_A4_vcmphgtui + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpw.eq(Rss32,#s8) + C Intrinsic Prototype: Byte Q6_p_vcmpw_eq_PI(Word64 Rss, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpw_eq_PI __builtin_HEXAGON_A4_vcmpweqi + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpw.gt(Rss32,#s8) + C Intrinsic Prototype: Byte Q6_p_vcmpw_gt_PI(Word64 Rss, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpw_gt_PI __builtin_HEXAGON_A4_vcmpwgti + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpw.gtu(Rss32,#u7) + C Intrinsic Prototype: Byte Q6_p_vcmpw_gtu_PI(Word64 Rss, Word32 Iu7) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpw_gtu_PI __builtin_HEXAGON_A4_vcmpwgtui + +/* ========================================================================== + Assembly Syntax: Rxx32=vrmaxh(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrmaxh_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmaxh_PR __builtin_HEXAGON_A4_vrmaxh + +/* ========================================================================== + Assembly Syntax: Rxx32=vrmaxuh(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrmaxuh_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmaxuh_PR __builtin_HEXAGON_A4_vrmaxuh + +/* ========================================================================== + Assembly Syntax: Rxx32=vrmaxuw(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrmaxuw_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmaxuw_PR __builtin_HEXAGON_A4_vrmaxuw + +/* ========================================================================== + Assembly Syntax: Rxx32=vrmaxw(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrmaxw_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmaxw_PR __builtin_HEXAGON_A4_vrmaxw + +/* ========================================================================== + Assembly Syntax: Rxx32=vrminh(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrminh_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrminh_PR __builtin_HEXAGON_A4_vrminh + +/* ========================================================================== + Assembly Syntax: Rxx32=vrminuh(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrminuh_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrminuh_PR __builtin_HEXAGON_A4_vrminuh + +/* ========================================================================== + Assembly Syntax: Rxx32=vrminuw(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrminuw_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrminuw_PR __builtin_HEXAGON_A4_vrminuw + +/* ========================================================================== + Assembly Syntax: Rxx32=vrminw(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrminw_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrminw_PR __builtin_HEXAGON_A4_vrminw + +/* ========================================================================== + Assembly Syntax: Rd32=vaddhub(Rss32,Rtt32):sat + C Intrinsic Prototype: Word32 Q6_R_vaddhub_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vaddhub_PP_sat __builtin_HEXAGON_A5_vaddhubs + +/* ========================================================================== + Assembly Syntax: Pd4=all8(Ps4) + C Intrinsic Prototype: Byte Q6_p_all8_p(Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_all8_p __builtin_HEXAGON_C2_all8 + +/* ========================================================================== + Assembly Syntax: Pd4=and(Pt4,Ps4) + C Intrinsic Prototype: Byte Q6_p_and_pp(Byte Pt, Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_and_pp __builtin_HEXAGON_C2_and + +/* ========================================================================== + Assembly Syntax: Pd4=and(Pt4,!Ps4) + C Intrinsic Prototype: Byte Q6_p_and_pnp(Byte Pt, Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_and_pnp __builtin_HEXAGON_C2_andn + +/* ========================================================================== + Assembly Syntax: Pd4=any8(Ps4) + C Intrinsic Prototype: Byte Q6_p_any8_p(Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_any8_p __builtin_HEXAGON_C2_any8 + +/* ========================================================================== + Assembly Syntax: Pd4=bitsclr(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_bitsclr_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_bitsclr_RR __builtin_HEXAGON_C2_bitsclr + +/* ========================================================================== + Assembly Syntax: Pd4=bitsclr(Rs32,#u6) + C Intrinsic Prototype: Byte Q6_p_bitsclr_RI(Word32 Rs, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_bitsclr_RI __builtin_HEXAGON_C2_bitsclri + +/* ========================================================================== + Assembly Syntax: Pd4=bitsset(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_bitsset_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_bitsset_RR __builtin_HEXAGON_C2_bitsset + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.eq(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmp_eq_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_eq_RR __builtin_HEXAGON_C2_cmpeq + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.eq(Rs32,#s10) + C Intrinsic Prototype: Byte Q6_p_cmp_eq_RI(Word32 Rs, Word32 Is10) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_eq_RI __builtin_HEXAGON_C2_cmpeqi + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.eq(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_cmp_eq_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmp_eq_PP __builtin_HEXAGON_C2_cmpeqp + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.ge(Rs32,#s8) + C Intrinsic Prototype: Byte Q6_p_cmp_ge_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_ge_RI __builtin_HEXAGON_C2_cmpgei + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.geu(Rs32,#u8) + C Intrinsic Prototype: Byte Q6_p_cmp_geu_RI(Word32 Rs, Word32 Iu8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_geu_RI __builtin_HEXAGON_C2_cmpgeui + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.gt(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmp_gt_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_gt_RR __builtin_HEXAGON_C2_cmpgt + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.gt(Rs32,#s10) + C Intrinsic Prototype: Byte Q6_p_cmp_gt_RI(Word32 Rs, Word32 Is10) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_gt_RI __builtin_HEXAGON_C2_cmpgti + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.gt(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_cmp_gt_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmp_gt_PP __builtin_HEXAGON_C2_cmpgtp + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.gtu(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmp_gtu_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_gtu_RR __builtin_HEXAGON_C2_cmpgtu + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.gtu(Rs32,#u9) + C Intrinsic Prototype: Byte Q6_p_cmp_gtu_RI(Word32 Rs, Word32 Iu9) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_gtu_RI __builtin_HEXAGON_C2_cmpgtui + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.gtu(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_cmp_gtu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmp_gtu_PP __builtin_HEXAGON_C2_cmpgtup + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.lt(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmp_lt_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_lt_RR __builtin_HEXAGON_C2_cmplt + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.ltu(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmp_ltu_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_ltu_RR __builtin_HEXAGON_C2_cmpltu + +/* ========================================================================== + Assembly Syntax: Rdd32=mask(Pt4) + C Intrinsic Prototype: Word64 Q6_P_mask_p(Byte Pt) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mask_p __builtin_HEXAGON_C2_mask + +/* ========================================================================== + Assembly Syntax: Rd32=mux(Pu4,Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mux_pRR(Byte Pu, Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_mux_pRR __builtin_HEXAGON_C2_mux + +/* ========================================================================== + Assembly Syntax: Rd32=mux(Pu4,#s8,#S8) + C Intrinsic Prototype: Word32 Q6_R_mux_pII(Byte Pu, Word32 Is8, Word32 IS8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_mux_pII __builtin_HEXAGON_C2_muxii + +/* ========================================================================== + Assembly Syntax: Rd32=mux(Pu4,Rs32,#s8) + C Intrinsic Prototype: Word32 Q6_R_mux_pRI(Byte Pu, Word32 Rs, Word32 Is8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_mux_pRI __builtin_HEXAGON_C2_muxir + +/* ========================================================================== + Assembly Syntax: Rd32=mux(Pu4,#s8,Rs32) + C Intrinsic Prototype: Word32 Q6_R_mux_pIR(Byte Pu, Word32 Is8, Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_mux_pIR __builtin_HEXAGON_C2_muxri + +/* ========================================================================== + Assembly Syntax: Pd4=not(Ps4) + C Intrinsic Prototype: Byte Q6_p_not_p(Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_p __builtin_HEXAGON_C2_not + +/* ========================================================================== + Assembly Syntax: Pd4=or(Pt4,Ps4) + C Intrinsic Prototype: Byte Q6_p_or_pp(Byte Pt, Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_or_pp __builtin_HEXAGON_C2_or + +/* ========================================================================== + Assembly Syntax: Pd4=or(Pt4,!Ps4) + C Intrinsic Prototype: Byte Q6_p_or_pnp(Byte Pt, Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_or_pnp __builtin_HEXAGON_C2_orn + +/* ========================================================================== + Assembly Syntax: Pd4=Ps4 + C Intrinsic Prototype: Byte Q6_p_equals_p(Byte Ps) + Instruction Type: MAPPING + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_equals_p __builtin_HEXAGON_C2_pxfer_map + +/* ========================================================================== + Assembly Syntax: Rd32=Ps4 + C Intrinsic Prototype: Word32 Q6_R_equals_p(Byte Ps) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_equals_p __builtin_HEXAGON_C2_tfrpr + +/* ========================================================================== + Assembly Syntax: Pd4=Rs32 + C Intrinsic Prototype: Byte Q6_p_equals_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_equals_R __builtin_HEXAGON_C2_tfrrp + +/* ========================================================================== + Assembly Syntax: Rd32=vitpack(Ps4,Pt4) + C Intrinsic Prototype: Word32 Q6_R_vitpack_pp(Byte Ps, Byte Pt) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vitpack_pp __builtin_HEXAGON_C2_vitpack + +/* ========================================================================== + Assembly Syntax: Rdd32=vmux(Pu4,Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vmux_pPP(Byte Pu, Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmux_pPP __builtin_HEXAGON_C2_vmux + +/* ========================================================================== + Assembly Syntax: Pd4=xor(Ps4,Pt4) + C Intrinsic Prototype: Byte Q6_p_xor_pp(Byte Ps, Byte Pt) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_xor_pp __builtin_HEXAGON_C2_xor + +/* ========================================================================== + Assembly Syntax: Pd4=and(Ps4,and(Pt4,Pu4)) + C Intrinsic Prototype: Byte Q6_p_and_and_ppp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_and_and_ppp __builtin_HEXAGON_C4_and_and + +/* ========================================================================== + Assembly Syntax: Pd4=and(Ps4,and(Pt4,!Pu4)) + C Intrinsic Prototype: Byte Q6_p_and_and_ppnp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_and_and_ppnp __builtin_HEXAGON_C4_and_andn + +/* ========================================================================== + Assembly Syntax: Pd4=and(Ps4,or(Pt4,Pu4)) + C Intrinsic Prototype: Byte Q6_p_and_or_ppp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_and_or_ppp __builtin_HEXAGON_C4_and_or + +/* ========================================================================== + Assembly Syntax: Pd4=and(Ps4,or(Pt4,!Pu4)) + C Intrinsic Prototype: Byte Q6_p_and_or_ppnp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_and_or_ppnp __builtin_HEXAGON_C4_and_orn + +/* ========================================================================== + Assembly Syntax: Pd4=!cmp.gt(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_not_cmp_gt_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_not_cmp_gt_RR __builtin_HEXAGON_C4_cmplte + +/* ========================================================================== + Assembly Syntax: Pd4=!cmp.gt(Rs32,#s10) + C Intrinsic Prototype: Byte Q6_p_not_cmp_gt_RI(Word32 Rs, Word32 Is10) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_not_cmp_gt_RI __builtin_HEXAGON_C4_cmpltei + +/* ========================================================================== + Assembly Syntax: Pd4=!cmp.gtu(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_not_cmp_gtu_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_not_cmp_gtu_RR __builtin_HEXAGON_C4_cmplteu + +/* ========================================================================== + Assembly Syntax: Pd4=!cmp.gtu(Rs32,#u9) + C Intrinsic Prototype: Byte Q6_p_not_cmp_gtu_RI(Word32 Rs, Word32 Iu9) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_not_cmp_gtu_RI __builtin_HEXAGON_C4_cmplteui + +/* ========================================================================== + Assembly Syntax: Pd4=!cmp.eq(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_not_cmp_eq_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_not_cmp_eq_RR __builtin_HEXAGON_C4_cmpneq + +/* ========================================================================== + Assembly Syntax: Pd4=!cmp.eq(Rs32,#s10) + C Intrinsic Prototype: Byte Q6_p_not_cmp_eq_RI(Word32 Rs, Word32 Is10) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_not_cmp_eq_RI __builtin_HEXAGON_C4_cmpneqi + +/* ========================================================================== + Assembly Syntax: Pd4=fastcorner9(Ps4,Pt4) + C Intrinsic Prototype: Byte Q6_p_fastcorner9_pp(Byte Ps, Byte Pt) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_fastcorner9_pp __builtin_HEXAGON_C4_fastcorner9 + +/* ========================================================================== + Assembly Syntax: Pd4=!fastcorner9(Ps4,Pt4) + C Intrinsic Prototype: Byte Q6_p_not_fastcorner9_pp(Byte Ps, Byte Pt) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_fastcorner9_pp __builtin_HEXAGON_C4_fastcorner9_not + +/* ========================================================================== + Assembly Syntax: Pd4=!bitsclr(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_not_bitsclr_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_bitsclr_RR __builtin_HEXAGON_C4_nbitsclr + +/* ========================================================================== + Assembly Syntax: Pd4=!bitsclr(Rs32,#u6) + C Intrinsic Prototype: Byte Q6_p_not_bitsclr_RI(Word32 Rs, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_bitsclr_RI __builtin_HEXAGON_C4_nbitsclri + +/* ========================================================================== + Assembly Syntax: Pd4=!bitsset(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_not_bitsset_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_bitsset_RR __builtin_HEXAGON_C4_nbitsset + +/* ========================================================================== + Assembly Syntax: Pd4=or(Ps4,and(Pt4,Pu4)) + C Intrinsic Prototype: Byte Q6_p_or_and_ppp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_or_and_ppp __builtin_HEXAGON_C4_or_and + +/* ========================================================================== + Assembly Syntax: Pd4=or(Ps4,and(Pt4,!Pu4)) + C Intrinsic Prototype: Byte Q6_p_or_and_ppnp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_or_and_ppnp __builtin_HEXAGON_C4_or_andn + +/* ========================================================================== + Assembly Syntax: Pd4=or(Ps4,or(Pt4,Pu4)) + C Intrinsic Prototype: Byte Q6_p_or_or_ppp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_or_or_ppp __builtin_HEXAGON_C4_or_or + +/* ========================================================================== + Assembly Syntax: Pd4=or(Ps4,or(Pt4,!Pu4)) + C Intrinsic Prototype: Byte Q6_p_or_or_ppnp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_or_or_ppnp __builtin_HEXAGON_C4_or_orn + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_d2df(Rss32) + C Intrinsic Prototype: Float64 Q6_P_convert_d2df_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_d2df_P __builtin_HEXAGON_F2_conv_d2df + +/* ========================================================================== + Assembly Syntax: Rd32=convert_d2sf(Rss32) + C Intrinsic Prototype: Float32 Q6_R_convert_d2sf_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_d2sf_P __builtin_HEXAGON_F2_conv_d2sf + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_df2d(Rss32) + C Intrinsic Prototype: Word64 Q6_P_convert_df2d_P(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_df2d_P __builtin_HEXAGON_F2_conv_df2d + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_df2d(Rss32):chop + C Intrinsic Prototype: Word64 Q6_P_convert_df2d_P_chop(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_df2d_P_chop __builtin_HEXAGON_F2_conv_df2d_chop + +/* ========================================================================== + Assembly Syntax: Rd32=convert_df2sf(Rss32) + C Intrinsic Prototype: Float32 Q6_R_convert_df2sf_P(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_df2sf_P __builtin_HEXAGON_F2_conv_df2sf + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_df2ud(Rss32) + C Intrinsic Prototype: Word64 Q6_P_convert_df2ud_P(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_df2ud_P __builtin_HEXAGON_F2_conv_df2ud + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_df2ud(Rss32):chop + C Intrinsic Prototype: Word64 Q6_P_convert_df2ud_P_chop(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_df2ud_P_chop __builtin_HEXAGON_F2_conv_df2ud_chop + +/* ========================================================================== + Assembly Syntax: Rd32=convert_df2uw(Rss32) + C Intrinsic Prototype: Word32 Q6_R_convert_df2uw_P(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_df2uw_P __builtin_HEXAGON_F2_conv_df2uw + +/* ========================================================================== + Assembly Syntax: Rd32=convert_df2uw(Rss32):chop + C Intrinsic Prototype: Word32 Q6_R_convert_df2uw_P_chop(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_df2uw_P_chop __builtin_HEXAGON_F2_conv_df2uw_chop + +/* ========================================================================== + Assembly Syntax: Rd32=convert_df2w(Rss32) + C Intrinsic Prototype: Word32 Q6_R_convert_df2w_P(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_df2w_P __builtin_HEXAGON_F2_conv_df2w + +/* ========================================================================== + Assembly Syntax: Rd32=convert_df2w(Rss32):chop + C Intrinsic Prototype: Word32 Q6_R_convert_df2w_P_chop(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_df2w_P_chop __builtin_HEXAGON_F2_conv_df2w_chop + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_sf2d(Rs32) + C Intrinsic Prototype: Word64 Q6_P_convert_sf2d_R(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_sf2d_R __builtin_HEXAGON_F2_conv_sf2d + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_sf2d(Rs32):chop + C Intrinsic Prototype: Word64 Q6_P_convert_sf2d_R_chop(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_sf2d_R_chop __builtin_HEXAGON_F2_conv_sf2d_chop + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_sf2df(Rs32) + C Intrinsic Prototype: Float64 Q6_P_convert_sf2df_R(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_sf2df_R __builtin_HEXAGON_F2_conv_sf2df + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_sf2ud(Rs32) + C Intrinsic Prototype: Word64 Q6_P_convert_sf2ud_R(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_sf2ud_R __builtin_HEXAGON_F2_conv_sf2ud + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_sf2ud(Rs32):chop + C Intrinsic Prototype: Word64 Q6_P_convert_sf2ud_R_chop(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_sf2ud_R_chop __builtin_HEXAGON_F2_conv_sf2ud_chop + +/* ========================================================================== + Assembly Syntax: Rd32=convert_sf2uw(Rs32) + C Intrinsic Prototype: Word32 Q6_R_convert_sf2uw_R(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_sf2uw_R __builtin_HEXAGON_F2_conv_sf2uw + +/* ========================================================================== + Assembly Syntax: Rd32=convert_sf2uw(Rs32):chop + C Intrinsic Prototype: Word32 Q6_R_convert_sf2uw_R_chop(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_sf2uw_R_chop __builtin_HEXAGON_F2_conv_sf2uw_chop + +/* ========================================================================== + Assembly Syntax: Rd32=convert_sf2w(Rs32) + C Intrinsic Prototype: Word32 Q6_R_convert_sf2w_R(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_sf2w_R __builtin_HEXAGON_F2_conv_sf2w + +/* ========================================================================== + Assembly Syntax: Rd32=convert_sf2w(Rs32):chop + C Intrinsic Prototype: Word32 Q6_R_convert_sf2w_R_chop(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_sf2w_R_chop __builtin_HEXAGON_F2_conv_sf2w_chop + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_ud2df(Rss32) + C Intrinsic Prototype: Float64 Q6_P_convert_ud2df_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_ud2df_P __builtin_HEXAGON_F2_conv_ud2df + +/* ========================================================================== + Assembly Syntax: Rd32=convert_ud2sf(Rss32) + C Intrinsic Prototype: Float32 Q6_R_convert_ud2sf_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_ud2sf_P __builtin_HEXAGON_F2_conv_ud2sf + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_uw2df(Rs32) + C Intrinsic Prototype: Float64 Q6_P_convert_uw2df_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_uw2df_R __builtin_HEXAGON_F2_conv_uw2df + +/* ========================================================================== + Assembly Syntax: Rd32=convert_uw2sf(Rs32) + C Intrinsic Prototype: Float32 Q6_R_convert_uw2sf_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_uw2sf_R __builtin_HEXAGON_F2_conv_uw2sf + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_w2df(Rs32) + C Intrinsic Prototype: Float64 Q6_P_convert_w2df_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_w2df_R __builtin_HEXAGON_F2_conv_w2df + +/* ========================================================================== + Assembly Syntax: Rd32=convert_w2sf(Rs32) + C Intrinsic Prototype: Float32 Q6_R_convert_w2sf_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_w2sf_R __builtin_HEXAGON_F2_conv_w2sf + +/* ========================================================================== + Assembly Syntax: Pd4=dfclass(Rss32,#u5) + C Intrinsic Prototype: Byte Q6_p_dfclass_PI(Float64 Rss, Word32 Iu5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_dfclass_PI __builtin_HEXAGON_F2_dfclass + +/* ========================================================================== + Assembly Syntax: Pd4=dfcmp.eq(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_dfcmp_eq_PP(Float64 Rss, Float64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_dfcmp_eq_PP __builtin_HEXAGON_F2_dfcmpeq + +/* ========================================================================== + Assembly Syntax: Pd4=dfcmp.ge(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_dfcmp_ge_PP(Float64 Rss, Float64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_dfcmp_ge_PP __builtin_HEXAGON_F2_dfcmpge + +/* ========================================================================== + Assembly Syntax: Pd4=dfcmp.gt(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_dfcmp_gt_PP(Float64 Rss, Float64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_dfcmp_gt_PP __builtin_HEXAGON_F2_dfcmpgt + +/* ========================================================================== + Assembly Syntax: Pd4=dfcmp.uo(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_dfcmp_uo_PP(Float64 Rss, Float64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_dfcmp_uo_PP __builtin_HEXAGON_F2_dfcmpuo + +/* ========================================================================== + Assembly Syntax: Rdd32=dfmake(#u10):neg + C Intrinsic Prototype: Float64 Q6_P_dfmake_I_neg(Word32 Iu10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmake_I_neg __builtin_HEXAGON_F2_dfimm_n + +/* ========================================================================== + Assembly Syntax: Rdd32=dfmake(#u10):pos + C Intrinsic Prototype: Float64 Q6_P_dfmake_I_pos(Word32 Iu10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmake_I_pos __builtin_HEXAGON_F2_dfimm_p + +/* ========================================================================== + Assembly Syntax: Rd32=sfadd(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfadd_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfadd_RR __builtin_HEXAGON_F2_sfadd + +/* ========================================================================== + Assembly Syntax: Pd4=sfclass(Rs32,#u5) + C Intrinsic Prototype: Byte Q6_p_sfclass_RI(Float32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_sfclass_RI __builtin_HEXAGON_F2_sfclass + +/* ========================================================================== + Assembly Syntax: Pd4=sfcmp.eq(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_sfcmp_eq_RR(Float32 Rs, Float32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_sfcmp_eq_RR __builtin_HEXAGON_F2_sfcmpeq + +/* ========================================================================== + Assembly Syntax: Pd4=sfcmp.ge(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_sfcmp_ge_RR(Float32 Rs, Float32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_sfcmp_ge_RR __builtin_HEXAGON_F2_sfcmpge + +/* ========================================================================== + Assembly Syntax: Pd4=sfcmp.gt(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_sfcmp_gt_RR(Float32 Rs, Float32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_sfcmp_gt_RR __builtin_HEXAGON_F2_sfcmpgt + +/* ========================================================================== + Assembly Syntax: Pd4=sfcmp.uo(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_sfcmp_uo_RR(Float32 Rs, Float32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_sfcmp_uo_RR __builtin_HEXAGON_F2_sfcmpuo + +/* ========================================================================== + Assembly Syntax: Rd32=sffixupd(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sffixupd_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sffixupd_RR __builtin_HEXAGON_F2_sffixupd + +/* ========================================================================== + Assembly Syntax: Rd32=sffixupn(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sffixupn_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sffixupn_RR __builtin_HEXAGON_F2_sffixupn + +/* ========================================================================== + Assembly Syntax: Rd32=sffixupr(Rs32) + C Intrinsic Prototype: Float32 Q6_R_sffixupr_R(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sffixupr_R __builtin_HEXAGON_F2_sffixupr + +/* ========================================================================== + Assembly Syntax: Rx32+=sfmpy(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfmpyacc_RR(Float32 Rx, Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmpyacc_RR __builtin_HEXAGON_F2_sffma + +/* ========================================================================== + Assembly Syntax: Rx32+=sfmpy(Rs32,Rt32):lib + C Intrinsic Prototype: Float32 Q6_R_sfmpyacc_RR_lib(Float32 Rx, Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmpyacc_RR_lib __builtin_HEXAGON_F2_sffma_lib + +/* ========================================================================== + Assembly Syntax: Rx32+=sfmpy(Rs32,Rt32,Pu4):scale + C Intrinsic Prototype: Float32 Q6_R_sfmpyacc_RRp_scale(Float32 Rx, Float32 Rs, Float32 Rt, Byte Pu) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmpyacc_RRp_scale __builtin_HEXAGON_F2_sffma_sc + +/* ========================================================================== + Assembly Syntax: Rx32-=sfmpy(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfmpynac_RR(Float32 Rx, Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmpynac_RR __builtin_HEXAGON_F2_sffms + +/* ========================================================================== + Assembly Syntax: Rx32-=sfmpy(Rs32,Rt32):lib + C Intrinsic Prototype: Float32 Q6_R_sfmpynac_RR_lib(Float32 Rx, Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmpynac_RR_lib __builtin_HEXAGON_F2_sffms_lib + +/* ========================================================================== + Assembly Syntax: Rd32=sfmake(#u10):neg + C Intrinsic Prototype: Float32 Q6_R_sfmake_I_neg(Word32 Iu10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmake_I_neg __builtin_HEXAGON_F2_sfimm_n + +/* ========================================================================== + Assembly Syntax: Rd32=sfmake(#u10):pos + C Intrinsic Prototype: Float32 Q6_R_sfmake_I_pos(Word32 Iu10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmake_I_pos __builtin_HEXAGON_F2_sfimm_p + +/* ========================================================================== + Assembly Syntax: Rd32=sfmax(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfmax_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmax_RR __builtin_HEXAGON_F2_sfmax + +/* ========================================================================== + Assembly Syntax: Rd32=sfmin(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfmin_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmin_RR __builtin_HEXAGON_F2_sfmin + +/* ========================================================================== + Assembly Syntax: Rd32=sfmpy(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfmpy_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmpy_RR __builtin_HEXAGON_F2_sfmpy + +/* ========================================================================== + Assembly Syntax: Rd32=sfsub(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfsub_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfsub_RR __builtin_HEXAGON_F2_sfsub + +/* ========================================================================== + Assembly Syntax: Rd32=memb(Rx32++#s4:0:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memb_IM_circ(void** Rx, Word32 Is4_0, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memb_IM_circ __builtin_HEXAGON_L2_loadrb_pci + +/* ========================================================================== + Assembly Syntax: Rd32=memb(Rx32++I:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memb_M_circ(void** Rx, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memb_M_circ __builtin_HEXAGON_L2_loadrb_pcr + +/* ========================================================================== + Assembly Syntax: Rdd32=memd(Rx32++#s4:3:circ(Mu2)) + C Intrinsic Prototype: Word64 Q6_P_memd_IM_circ(void** Rx, Word32 Is4_3, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_P_memd_IM_circ __builtin_HEXAGON_L2_loadrd_pci + +/* ========================================================================== + Assembly Syntax: Rdd32=memd(Rx32++I:circ(Mu2)) + C Intrinsic Prototype: Word64 Q6_P_memd_M_circ(void** Rx, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_P_memd_M_circ __builtin_HEXAGON_L2_loadrd_pcr + +/* ========================================================================== + Assembly Syntax: Rd32=memh(Rx32++#s4:1:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memh_IM_circ(void** Rx, Word32 Is4_1, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memh_IM_circ __builtin_HEXAGON_L2_loadrh_pci + +/* ========================================================================== + Assembly Syntax: Rd32=memh(Rx32++I:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memh_M_circ(void** Rx, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memh_M_circ __builtin_HEXAGON_L2_loadrh_pcr + +/* ========================================================================== + Assembly Syntax: Rd32=memw(Rx32++#s4:2:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memw_IM_circ(void** Rx, Word32 Is4_2, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memw_IM_circ __builtin_HEXAGON_L2_loadri_pci + +/* ========================================================================== + Assembly Syntax: Rd32=memw(Rx32++I:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memw_M_circ(void** Rx, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memw_M_circ __builtin_HEXAGON_L2_loadri_pcr + +/* ========================================================================== + Assembly Syntax: Rd32=memub(Rx32++#s4:0:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memub_IM_circ(void** Rx, Word32 Is4_0, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memub_IM_circ __builtin_HEXAGON_L2_loadrub_pci + +/* ========================================================================== + Assembly Syntax: Rd32=memub(Rx32++I:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memub_M_circ(void** Rx, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memub_M_circ __builtin_HEXAGON_L2_loadrub_pcr + +/* ========================================================================== + Assembly Syntax: Rd32=memuh(Rx32++#s4:1:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memuh_IM_circ(void** Rx, Word32 Is4_1, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memuh_IM_circ __builtin_HEXAGON_L2_loadruh_pci + +/* ========================================================================== + Assembly Syntax: Rd32=memuh(Rx32++I:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memuh_M_circ(void** Rx, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memuh_M_circ __builtin_HEXAGON_L2_loadruh_pcr + +/* ========================================================================== + Assembly Syntax: Rx32+=add(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_addacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_addacc_RR __builtin_HEXAGON_M2_acci + +/* ========================================================================== + Assembly Syntax: Rx32+=add(Rs32,#s8) + C Intrinsic Prototype: Word32 Q6_R_addacc_RI(Word32 Rx, Word32 Rs, Word32 Is8) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_addacc_RI __builtin_HEXAGON_M2_accii + +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpyi(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyiacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyiacc_RR __builtin_HEXAGON_M2_cmaci_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpyr(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyracc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyracc_RR __builtin_HEXAGON_M2_cmacr_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpy(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_cmpyacc_RR_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyacc_RR_sat __builtin_HEXAGON_M2_cmacs_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpy(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_cmpyacc_RR_s1_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyacc_RR_s1_sat __builtin_HEXAGON_M2_cmacs_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpy(Rs32,Rt32*):sat + C Intrinsic Prototype: Word64 Q6_P_cmpyacc_RR_conj_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyacc_RR_conj_sat __builtin_HEXAGON_M2_cmacsc_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpy(Rs32,Rt32*):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_cmpyacc_RR_conj_s1_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyacc_RR_conj_s1_sat __builtin_HEXAGON_M2_cmacsc_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=cmpyi(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyi_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyi_RR __builtin_HEXAGON_M2_cmpyi_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=cmpyr(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyr_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyr_RR __builtin_HEXAGON_M2_cmpyr_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=cmpy(Rs32,Rt32):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpy_RR_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpy_RR_rnd_sat __builtin_HEXAGON_M2_cmpyrs_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=cmpy(Rs32,Rt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpy_RR_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpy_RR_s1_rnd_sat __builtin_HEXAGON_M2_cmpyrs_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=cmpy(Rs32,Rt32*):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpy_RR_conj_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpy_RR_conj_rnd_sat __builtin_HEXAGON_M2_cmpyrsc_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=cmpy(Rs32,Rt32*):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpy_RR_conj_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpy_RR_conj_s1_rnd_sat __builtin_HEXAGON_M2_cmpyrsc_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=cmpy(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_cmpy_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpy_RR_sat __builtin_HEXAGON_M2_cmpys_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=cmpy(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_cmpy_RR_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpy_RR_s1_sat __builtin_HEXAGON_M2_cmpys_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=cmpy(Rs32,Rt32*):sat + C Intrinsic Prototype: Word64 Q6_P_cmpy_RR_conj_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpy_RR_conj_sat __builtin_HEXAGON_M2_cmpysc_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=cmpy(Rs32,Rt32*):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_cmpy_RR_conj_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpy_RR_conj_s1_sat __builtin_HEXAGON_M2_cmpysc_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=cmpy(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_cmpynac_RR_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpynac_RR_sat __builtin_HEXAGON_M2_cnacs_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=cmpy(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_cmpynac_RR_s1_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpynac_RR_s1_sat __builtin_HEXAGON_M2_cnacs_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=cmpy(Rs32,Rt32*):sat + C Intrinsic Prototype: Word64 Q6_P_cmpynac_RR_conj_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpynac_RR_conj_sat __builtin_HEXAGON_M2_cnacsc_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=cmpy(Rs32,Rt32*):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_cmpynac_RR_conj_s1_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpynac_RR_conj_s1_sat __builtin_HEXAGON_M2_cnacsc_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RR __builtin_HEXAGON_M2_dpmpyss_acc_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_mpynac_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RR __builtin_HEXAGON_M2_dpmpyss_nac_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32):rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RR_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RR_rnd __builtin_HEXAGON_M2_dpmpyss_rnd_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_mpy_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RR __builtin_HEXAGON_M2_dpmpyss_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RR __builtin_HEXAGON_M2_dpmpyuu_acc_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RR __builtin_HEXAGON_M2_dpmpyuu_nac_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32,Rt32) + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RR __builtin_HEXAGON_M2_dpmpyuu_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32.h):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RRh_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RRh_s1_rnd_sat __builtin_HEXAGON_M2_hmmpyh_rs1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RRh_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RRh_s1_sat __builtin_HEXAGON_M2_hmmpyh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32.l):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RRl_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RRl_s1_rnd_sat __builtin_HEXAGON_M2_hmmpyl_rs1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RRl_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RRl_s1_sat __builtin_HEXAGON_M2_hmmpyl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyi(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mpyiacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyiacc_RR __builtin_HEXAGON_M2_maci + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyi(Rs32,#u8) + C Intrinsic Prototype: Word32 Q6_R_mpyinac_RI(Word32 Rx, Word32 Rs, Word32 Iu8) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyinac_RI __builtin_HEXAGON_M2_macsin + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyi(Rs32,#u8) + C Intrinsic Prototype: Word32 Q6_R_mpyiacc_RI(Word32 Rx, Word32 Rs, Word32 Iu8) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyiacc_RI __builtin_HEXAGON_M2_macsip + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywoh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywohacc_PP_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywohacc_PP_rnd_sat __builtin_HEXAGON_M2_mmachs_rs0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywoh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywohacc_PP_s1_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywohacc_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmachs_rs1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywoh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpywohacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywohacc_PP_sat __builtin_HEXAGON_M2_mmachs_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywoh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywohacc_PP_s1_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywohacc_PP_s1_sat __builtin_HEXAGON_M2_mmachs_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywehacc_PP_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywehacc_PP_rnd_sat __builtin_HEXAGON_M2_mmacls_rs0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywehacc_PP_s1_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywehacc_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmacls_rs1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpywehacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywehacc_PP_sat __builtin_HEXAGON_M2_mmacls_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywehacc_PP_s1_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywehacc_PP_s1_sat __builtin_HEXAGON_M2_mmacls_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywouh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouhacc_PP_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouhacc_PP_rnd_sat __builtin_HEXAGON_M2_mmacuhs_rs0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywouh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouhacc_PP_s1_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouhacc_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmacuhs_rs1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywouh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouhacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouhacc_PP_sat __builtin_HEXAGON_M2_mmacuhs_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywouh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouhacc_PP_s1_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouhacc_PP_s1_sat __builtin_HEXAGON_M2_mmacuhs_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweuh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuhacc_PP_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuhacc_PP_rnd_sat __builtin_HEXAGON_M2_mmaculs_rs0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweuh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuhacc_PP_s1_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuhacc_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmaculs_rs1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweuh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuhacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuhacc_PP_sat __builtin_HEXAGON_M2_mmaculs_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweuh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuhacc_PP_s1_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuhacc_PP_s1_sat __builtin_HEXAGON_M2_mmaculs_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywoh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywoh_PP_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywoh_PP_rnd_sat __builtin_HEXAGON_M2_mmpyh_rs0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywoh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywoh_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywoh_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmpyh_rs1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywoh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpywoh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywoh_PP_sat __builtin_HEXAGON_M2_mmpyh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywoh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywoh_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywoh_PP_s1_sat __builtin_HEXAGON_M2_mmpyh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweh_PP_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweh_PP_rnd_sat __builtin_HEXAGON_M2_mmpyl_rs0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweh_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweh_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmpyl_rs1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweh_PP_sat __builtin_HEXAGON_M2_mmpyl_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweh_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweh_PP_s1_sat __builtin_HEXAGON_M2_mmpyl_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywouh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouh_PP_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouh_PP_rnd_sat __builtin_HEXAGON_M2_mmpyuh_rs0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywouh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouh_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouh_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmpyuh_rs1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywouh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouh_PP_sat __builtin_HEXAGON_M2_mmpyuh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywouh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouh_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouh_PP_s1_sat __builtin_HEXAGON_M2_mmpyuh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweuh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuh_PP_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuh_PP_rnd_sat __builtin_HEXAGON_M2_mmpyul_rs0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweuh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuh_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuh_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmpyul_rs1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweuh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuh_PP_sat __builtin_HEXAGON_M2_mmpyul_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweuh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuh_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuh_PP_s1_sat __builtin_HEXAGON_M2_mmpyul_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRh __builtin_HEXAGON_M2_mpy_acc_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRh_s1 __builtin_HEXAGON_M2_mpy_acc_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRl __builtin_HEXAGON_M2_mpy_acc_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRl_s1 __builtin_HEXAGON_M2_mpy_acc_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRh __builtin_HEXAGON_M2_mpy_acc_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRh_s1 __builtin_HEXAGON_M2_mpy_acc_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRl __builtin_HEXAGON_M2_mpy_acc_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRl_s1 __builtin_HEXAGON_M2_mpy_acc_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.h):sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRh_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRh_sat __builtin_HEXAGON_M2_mpy_acc_sat_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRh_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRh_s1_sat __builtin_HEXAGON_M2_mpy_acc_sat_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.l):sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRl_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRl_sat __builtin_HEXAGON_M2_mpy_acc_sat_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRl_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRl_s1_sat __builtin_HEXAGON_M2_mpy_acc_sat_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.h):sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRh_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRh_sat __builtin_HEXAGON_M2_mpy_acc_sat_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRh_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRh_s1_sat __builtin_HEXAGON_M2_mpy_acc_sat_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.l):sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRl_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRl_sat __builtin_HEXAGON_M2_mpy_acc_sat_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRl_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRl_s1_sat __builtin_HEXAGON_M2_mpy_acc_sat_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh __builtin_HEXAGON_M2_mpy_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_s1 __builtin_HEXAGON_M2_mpy_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl __builtin_HEXAGON_M2_mpy_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_s1 __builtin_HEXAGON_M2_mpy_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh __builtin_HEXAGON_M2_mpy_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_s1 __builtin_HEXAGON_M2_mpy_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl __builtin_HEXAGON_M2_mpy_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_s1 __builtin_HEXAGON_M2_mpy_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRh __builtin_HEXAGON_M2_mpy_nac_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRh_s1 __builtin_HEXAGON_M2_mpy_nac_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRl __builtin_HEXAGON_M2_mpy_nac_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRl_s1 __builtin_HEXAGON_M2_mpy_nac_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRh __builtin_HEXAGON_M2_mpy_nac_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRh_s1 __builtin_HEXAGON_M2_mpy_nac_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRl __builtin_HEXAGON_M2_mpy_nac_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRl_s1 __builtin_HEXAGON_M2_mpy_nac_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.h):sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRh_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRh_sat __builtin_HEXAGON_M2_mpy_nac_sat_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRh_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRh_s1_sat __builtin_HEXAGON_M2_mpy_nac_sat_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.l):sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRl_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRl_sat __builtin_HEXAGON_M2_mpy_nac_sat_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRl_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRl_s1_sat __builtin_HEXAGON_M2_mpy_nac_sat_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.h):sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRh_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRh_sat __builtin_HEXAGON_M2_mpy_nac_sat_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRh_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRh_s1_sat __builtin_HEXAGON_M2_mpy_nac_sat_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.l):sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRl_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRl_sat __builtin_HEXAGON_M2_mpy_nac_sat_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRl_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRl_s1_sat __builtin_HEXAGON_M2_mpy_nac_sat_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_rnd __builtin_HEXAGON_M2_mpy_rnd_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):<<1:rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_s1_rnd __builtin_HEXAGON_M2_mpy_rnd_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_rnd __builtin_HEXAGON_M2_mpy_rnd_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):<<1:rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_s1_rnd __builtin_HEXAGON_M2_mpy_rnd_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_rnd __builtin_HEXAGON_M2_mpy_rnd_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):<<1:rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_s1_rnd __builtin_HEXAGON_M2_mpy_rnd_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_rnd __builtin_HEXAGON_M2_mpy_rnd_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):<<1:rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_s1_rnd __builtin_HEXAGON_M2_mpy_rnd_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_sat __builtin_HEXAGON_M2_mpy_sat_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_s1_sat __builtin_HEXAGON_M2_mpy_sat_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_sat __builtin_HEXAGON_M2_mpy_sat_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_s1_sat __builtin_HEXAGON_M2_mpy_sat_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_sat __builtin_HEXAGON_M2_mpy_sat_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_s1_sat __builtin_HEXAGON_M2_mpy_sat_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_sat __builtin_HEXAGON_M2_mpy_sat_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_s1_sat __builtin_HEXAGON_M2_mpy_sat_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_s1_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_s1_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_s1_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_s1_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mpy_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RR __builtin_HEXAGON_M2_mpy_up + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpy_RR_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RR_s1 __builtin_HEXAGON_M2_mpy_up_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RR_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RR_s1_sat __builtin_HEXAGON_M2_mpy_up_s1_sat + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RhRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RhRh __builtin_HEXAGON_M2_mpyd_acc_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RhRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RhRh_s1 __builtin_HEXAGON_M2_mpyd_acc_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RhRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RhRl __builtin_HEXAGON_M2_mpyd_acc_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RhRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RhRl_s1 __builtin_HEXAGON_M2_mpyd_acc_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RlRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RlRh __builtin_HEXAGON_M2_mpyd_acc_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RlRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RlRh_s1 __builtin_HEXAGON_M2_mpyd_acc_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RlRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RlRl __builtin_HEXAGON_M2_mpyd_acc_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RlRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RlRl_s1 __builtin_HEXAGON_M2_mpyd_acc_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRh __builtin_HEXAGON_M2_mpyd_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRh_s1 __builtin_HEXAGON_M2_mpyd_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRl __builtin_HEXAGON_M2_mpyd_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRl_s1 __builtin_HEXAGON_M2_mpyd_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRh __builtin_HEXAGON_M2_mpyd_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRh_s1 __builtin_HEXAGON_M2_mpyd_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRl __builtin_HEXAGON_M2_mpyd_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRl_s1 __builtin_HEXAGON_M2_mpyd_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpynac_RhRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RhRh __builtin_HEXAGON_M2_mpyd_nac_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpynac_RhRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RhRh_s1 __builtin_HEXAGON_M2_mpyd_nac_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpynac_RhRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RhRl __builtin_HEXAGON_M2_mpyd_nac_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpynac_RhRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RhRl_s1 __builtin_HEXAGON_M2_mpyd_nac_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpynac_RlRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RlRh __builtin_HEXAGON_M2_mpyd_nac_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpynac_RlRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RlRh_s1 __builtin_HEXAGON_M2_mpyd_nac_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpynac_RlRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RlRl __builtin_HEXAGON_M2_mpyd_nac_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpynac_RlRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RlRl_s1 __builtin_HEXAGON_M2_mpyd_nac_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.h):rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRh_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRh_rnd __builtin_HEXAGON_M2_mpyd_rnd_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.h):<<1:rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRh_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRh_s1_rnd __builtin_HEXAGON_M2_mpyd_rnd_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.l):rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRl_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRl_rnd __builtin_HEXAGON_M2_mpyd_rnd_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.l):<<1:rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRl_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRl_s1_rnd __builtin_HEXAGON_M2_mpyd_rnd_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.h):rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRh_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRh_rnd __builtin_HEXAGON_M2_mpyd_rnd_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.h):<<1:rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRh_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRh_s1_rnd __builtin_HEXAGON_M2_mpyd_rnd_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.l):rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRl_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRl_rnd __builtin_HEXAGON_M2_mpyd_rnd_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.l):<<1:rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRl_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRl_s1_rnd __builtin_HEXAGON_M2_mpyd_rnd_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyi(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mpyi_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyi_RR __builtin_HEXAGON_M2_mpyi + +/* ========================================================================== + Assembly Syntax: Rd32=mpyi(Rs32,#m9) + C Intrinsic Prototype: Word32 Q6_R_mpyi_RI(Word32 Rs, Word32 Im9) + Instruction Type: M + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_mpyi_RI __builtin_HEXAGON_M2_mpysmi + +/* ========================================================================== + Assembly Syntax: Rd32=mpysu(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mpysu_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpysu_RR __builtin_HEXAGON_M2_mpysu_up + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RhRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RhRh __builtin_HEXAGON_M2_mpyu_acc_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RhRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RhRh_s1 __builtin_HEXAGON_M2_mpyu_acc_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RhRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RhRl __builtin_HEXAGON_M2_mpyu_acc_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RhRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RhRl_s1 __builtin_HEXAGON_M2_mpyu_acc_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RlRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RlRh __builtin_HEXAGON_M2_mpyu_acc_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RlRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RlRh_s1 __builtin_HEXAGON_M2_mpyu_acc_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RlRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RlRl __builtin_HEXAGON_M2_mpyu_acc_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RlRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RlRl_s1 __builtin_HEXAGON_M2_mpyu_acc_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.h,Rt32.h) + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RhRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RhRh __builtin_HEXAGON_M2_mpyu_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RhRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RhRh_s1 __builtin_HEXAGON_M2_mpyu_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.h,Rt32.l) + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RhRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RhRl __builtin_HEXAGON_M2_mpyu_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RhRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RhRl_s1 __builtin_HEXAGON_M2_mpyu_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.l,Rt32.h) + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RlRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RlRh __builtin_HEXAGON_M2_mpyu_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RlRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RlRh_s1 __builtin_HEXAGON_M2_mpyu_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.l,Rt32.l) + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RlRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RlRl __builtin_HEXAGON_M2_mpyu_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RlRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RlRl_s1 __builtin_HEXAGON_M2_mpyu_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RhRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RhRh __builtin_HEXAGON_M2_mpyu_nac_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RhRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RhRh_s1 __builtin_HEXAGON_M2_mpyu_nac_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RhRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RhRl __builtin_HEXAGON_M2_mpyu_nac_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RhRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RhRl_s1 __builtin_HEXAGON_M2_mpyu_nac_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RlRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RlRh __builtin_HEXAGON_M2_mpyu_nac_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RlRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RlRh_s1 __builtin_HEXAGON_M2_mpyu_nac_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RlRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RlRl __builtin_HEXAGON_M2_mpyu_nac_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RlRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RlRl_s1 __builtin_HEXAGON_M2_mpyu_nac_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32,Rt32) + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RR __builtin_HEXAGON_M2_mpyu_up + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RhRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RhRh __builtin_HEXAGON_M2_mpyud_acc_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RhRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RhRh_s1 __builtin_HEXAGON_M2_mpyud_acc_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RhRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RhRl __builtin_HEXAGON_M2_mpyud_acc_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RhRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RhRl_s1 __builtin_HEXAGON_M2_mpyud_acc_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RlRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RlRh __builtin_HEXAGON_M2_mpyud_acc_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RlRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RlRh_s1 __builtin_HEXAGON_M2_mpyud_acc_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RlRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RlRl __builtin_HEXAGON_M2_mpyud_acc_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RlRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RlRl_s1 __builtin_HEXAGON_M2_mpyud_acc_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.h,Rt32.h) + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RhRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RhRh __builtin_HEXAGON_M2_mpyud_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RhRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RhRh_s1 __builtin_HEXAGON_M2_mpyud_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.h,Rt32.l) + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RhRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RhRl __builtin_HEXAGON_M2_mpyud_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RhRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RhRl_s1 __builtin_HEXAGON_M2_mpyud_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.l,Rt32.h) + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RlRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RlRh __builtin_HEXAGON_M2_mpyud_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RlRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RlRh_s1 __builtin_HEXAGON_M2_mpyud_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.l,Rt32.l) + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RlRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RlRl __builtin_HEXAGON_M2_mpyud_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RlRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RlRl_s1 __builtin_HEXAGON_M2_mpyud_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RhRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RhRh __builtin_HEXAGON_M2_mpyud_nac_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RhRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RhRh_s1 __builtin_HEXAGON_M2_mpyud_nac_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RhRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RhRl __builtin_HEXAGON_M2_mpyud_nac_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RhRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RhRl_s1 __builtin_HEXAGON_M2_mpyud_nac_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RlRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RlRh __builtin_HEXAGON_M2_mpyud_nac_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RlRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RlRh_s1 __builtin_HEXAGON_M2_mpyud_nac_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RlRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RlRl __builtin_HEXAGON_M2_mpyud_nac_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RlRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RlRl_s1 __builtin_HEXAGON_M2_mpyud_nac_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyui(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mpyui_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_mpyui_RR __builtin_HEXAGON_M2_mpyui + +/* ========================================================================== + Assembly Syntax: Rx32-=add(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_addnac_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_addnac_RR __builtin_HEXAGON_M2_nacci + +/* ========================================================================== + Assembly Syntax: Rx32-=add(Rs32,#s8) + C Intrinsic Prototype: Word32 Q6_R_addnac_RI(Word32 Rx, Word32 Rs, Word32 Is8) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_addnac_RI __builtin_HEXAGON_M2_naccii + +/* ========================================================================== + Assembly Syntax: Rx32+=sub(Rt32,Rs32) + C Intrinsic Prototype: Word32 Q6_R_subacc_RR(Word32 Rx, Word32 Rt, Word32 Rs) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_subacc_RR __builtin_HEXAGON_M2_subacc + +/* ========================================================================== + Assembly Syntax: Rdd32=vabsdiffh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vabsdiffh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsdiffh_PP __builtin_HEXAGON_M2_vabsdiffh + +/* ========================================================================== + Assembly Syntax: Rdd32=vabsdiffw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vabsdiffw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsdiffw_PP __builtin_HEXAGON_M2_vabsdiffw + +/* ========================================================================== + Assembly Syntax: Rxx32+=vcmpyi(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vcmpyiacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcmpyiacc_PP_sat __builtin_HEXAGON_M2_vcmac_s0_sat_i + +/* ========================================================================== + Assembly Syntax: Rxx32+=vcmpyr(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vcmpyracc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcmpyracc_PP_sat __builtin_HEXAGON_M2_vcmac_s0_sat_r + +/* ========================================================================== + Assembly Syntax: Rdd32=vcmpyi(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vcmpyi_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcmpyi_PP_sat __builtin_HEXAGON_M2_vcmpy_s0_sat_i + +/* ========================================================================== + Assembly Syntax: Rdd32=vcmpyr(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vcmpyr_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcmpyr_PP_sat __builtin_HEXAGON_M2_vcmpy_s0_sat_r + +/* ========================================================================== + Assembly Syntax: Rdd32=vcmpyi(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vcmpyi_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcmpyi_PP_s1_sat __builtin_HEXAGON_M2_vcmpy_s1_sat_i + +/* ========================================================================== + Assembly Syntax: Rdd32=vcmpyr(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vcmpyr_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcmpyr_PP_s1_sat __builtin_HEXAGON_M2_vcmpy_s1_sat_r + +/* ========================================================================== + Assembly Syntax: Rxx32+=vdmpy(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vdmpyacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vdmpyacc_PP_sat __builtin_HEXAGON_M2_vdmacs_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vdmpy(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vdmpyacc_PP_s1_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vdmpyacc_PP_s1_sat __builtin_HEXAGON_M2_vdmacs_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=vdmpy(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_vdmpy_PP_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vdmpy_PP_rnd_sat __builtin_HEXAGON_M2_vdmpyrs_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=vdmpy(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_vdmpy_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vdmpy_PP_s1_rnd_sat __builtin_HEXAGON_M2_vdmpyrs_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vdmpy(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vdmpy_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vdmpy_PP_sat __builtin_HEXAGON_M2_vdmpys_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vdmpy(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vdmpy_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vdmpy_PP_s1_sat __builtin_HEXAGON_M2_vdmpys_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyh(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vmpyhacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhacc_RR __builtin_HEXAGON_M2_vmac2 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyeh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vmpyehacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyehacc_PP __builtin_HEXAGON_M2_vmac2es + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyeh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyehacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyehacc_PP_sat __builtin_HEXAGON_M2_vmac2es_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyeh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyehacc_PP_s1_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyehacc_PP_s1_sat __builtin_HEXAGON_M2_vmac2es_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyh(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyhacc_RR_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhacc_RR_sat __builtin_HEXAGON_M2_vmac2s_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyh(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyhacc_RR_s1_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhacc_RR_s1_sat __builtin_HEXAGON_M2_vmac2s_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyhsu(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyhsuacc_RR_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhsuacc_RR_sat __builtin_HEXAGON_M2_vmac2su_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyhsu(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyhsuacc_RR_s1_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhsuacc_RR_s1_sat __builtin_HEXAGON_M2_vmac2su_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyeh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyeh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyeh_PP_sat __builtin_HEXAGON_M2_vmpy2es_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyeh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyeh_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyeh_PP_s1_sat __builtin_HEXAGON_M2_vmpy2es_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyh(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyh_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyh_RR_sat __builtin_HEXAGON_M2_vmpy2s_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=vmpyh(Rs32,Rt32):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_vmpyh_RR_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vmpyh_RR_rnd_sat __builtin_HEXAGON_M2_vmpy2s_s0pack + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyh(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyh_RR_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyh_RR_s1_sat __builtin_HEXAGON_M2_vmpy2s_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=vmpyh(Rs32,Rt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_vmpyh_RR_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vmpyh_RR_s1_rnd_sat __builtin_HEXAGON_M2_vmpy2s_s1pack + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyhsu(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyhsu_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhsu_RR_sat __builtin_HEXAGON_M2_vmpy2su_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyhsu(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyhsu_RR_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhsu_RR_s1_sat __builtin_HEXAGON_M2_vmpy2su_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=vraddh(Rss32,Rtt32) + C Intrinsic Prototype: Word32 Q6_R_vraddh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vraddh_PP __builtin_HEXAGON_M2_vraddh + +/* ========================================================================== + Assembly Syntax: Rd32=vradduh(Rss32,Rtt32) + C Intrinsic Prototype: Word32 Q6_R_vradduh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vradduh_PP __builtin_HEXAGON_M2_vradduh + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcmpyi(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyiacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyiacc_PP __builtin_HEXAGON_M2_vrcmaci_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcmpyi(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyiacc_PP_conj(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyiacc_PP_conj __builtin_HEXAGON_M2_vrcmaci_s0c + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcmpyr(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyracc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyracc_PP __builtin_HEXAGON_M2_vrcmacr_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcmpyr(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyracc_PP_conj(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyracc_PP_conj __builtin_HEXAGON_M2_vrcmacr_s0c + +/* ========================================================================== + Assembly Syntax: Rdd32=vrcmpyi(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyi_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyi_PP __builtin_HEXAGON_M2_vrcmpyi_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrcmpyi(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyi_PP_conj(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyi_PP_conj __builtin_HEXAGON_M2_vrcmpyi_s0c + +/* ========================================================================== + Assembly Syntax: Rdd32=vrcmpyr(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyr_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyr_PP __builtin_HEXAGON_M2_vrcmpyr_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrcmpyr(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyr_PP_conj(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyr_PP_conj __builtin_HEXAGON_M2_vrcmpyr_s0c + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcmpys(Rss32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vrcmpysacc_PR_s1_sat(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_vrcmpysacc_PR_s1_sat __builtin_HEXAGON_M2_vrcmpys_acc_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrcmpys(Rss32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vrcmpys_PR_s1_sat(Word64 Rss, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_vrcmpys_PR_s1_sat __builtin_HEXAGON_M2_vrcmpys_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=vrcmpys(Rss32,Rt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_vrcmpys_PR_s1_rnd_sat(Word64 Rss, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vrcmpys_PR_s1_rnd_sat __builtin_HEXAGON_M2_vrcmpys_s1rp + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpyh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpyhacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpyhacc_PP __builtin_HEXAGON_M2_vrmac_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpyh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpyh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpyh_PP __builtin_HEXAGON_M2_vrmpy_s0 + +/* ========================================================================== + Assembly Syntax: Rx32^=xor(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_xorxacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_xorxacc_RR __builtin_HEXAGON_M2_xor_xacc + +/* ========================================================================== + Assembly Syntax: Rx32&=and(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_andand_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andand_RR __builtin_HEXAGON_M4_and_and + +/* ========================================================================== + Assembly Syntax: Rx32&=and(Rs32,~Rt32) + C Intrinsic Prototype: Word32 Q6_R_andand_RnR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andand_RnR __builtin_HEXAGON_M4_and_andn + +/* ========================================================================== + Assembly Syntax: Rx32&=or(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_orand_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_orand_RR __builtin_HEXAGON_M4_and_or + +/* ========================================================================== + Assembly Syntax: Rx32&=xor(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_xorand_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_xorand_RR __builtin_HEXAGON_M4_and_xor + +/* ========================================================================== + Assembly Syntax: Rd32=cmpyiwh(Rss32,Rt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyiwh_PR_s1_rnd_sat(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpyiwh_PR_s1_rnd_sat __builtin_HEXAGON_M4_cmpyi_wh + +/* ========================================================================== + Assembly Syntax: Rd32=cmpyiwh(Rss32,Rt32*):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyiwh_PR_conj_s1_rnd_sat(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpyiwh_PR_conj_s1_rnd_sat __builtin_HEXAGON_M4_cmpyi_whc + +/* ========================================================================== + Assembly Syntax: Rd32=cmpyrwh(Rss32,Rt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyrwh_PR_s1_rnd_sat(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpyrwh_PR_s1_rnd_sat __builtin_HEXAGON_M4_cmpyr_wh + +/* ========================================================================== + Assembly Syntax: Rd32=cmpyrwh(Rss32,Rt32*):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyrwh_PR_conj_s1_rnd_sat(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpyrwh_PR_conj_s1_rnd_sat __builtin_HEXAGON_M4_cmpyr_whc + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RR_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RR_s1_sat __builtin_HEXAGON_M4_mac_up_s1_sat + +/* ========================================================================== + Assembly Syntax: Rd32=add(#u6,mpyi(Rs32,#U6)) + C Intrinsic Prototype: Word32 Q6_R_add_mpyi_IRI(Word32 Iu6, Word32 Rs, Word32 IU6) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_mpyi_IRI __builtin_HEXAGON_M4_mpyri_addi + +/* ========================================================================== + Assembly Syntax: Rd32=add(Ru32,mpyi(Rs32,#u6)) + C Intrinsic Prototype: Word32 Q6_R_add_mpyi_RRI(Word32 Ru, Word32 Rs, Word32 Iu6) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_mpyi_RRI __builtin_HEXAGON_M4_mpyri_addr + +/* ========================================================================== + Assembly Syntax: Rd32=add(Ru32,mpyi(#u6:2,Rs32)) + C Intrinsic Prototype: Word32 Q6_R_add_mpyi_RIR(Word32 Ru, Word32 Iu6_2, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_mpyi_RIR __builtin_HEXAGON_M4_mpyri_addr_u2 + +/* ========================================================================== + Assembly Syntax: Rd32=add(#u6,mpyi(Rs32,Rt32)) + C Intrinsic Prototype: Word32 Q6_R_add_mpyi_IRR(Word32 Iu6, Word32 Rs, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_mpyi_IRR __builtin_HEXAGON_M4_mpyrr_addi + +/* ========================================================================== + Assembly Syntax: Ry32=add(Ru32,mpyi(Ry32,Rs32)) + C Intrinsic Prototype: Word32 Q6_R_add_mpyi_RRR(Word32 Ru, Word32 Ry, Word32 Rs) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_mpyi_RRR __builtin_HEXAGON_M4_mpyrr_addr + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RR_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RR_s1_sat __builtin_HEXAGON_M4_nac_up_s1_sat + +/* ========================================================================== + Assembly Syntax: Rx32|=and(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_andor_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andor_RR __builtin_HEXAGON_M4_or_and + +/* ========================================================================== + Assembly Syntax: Rx32|=and(Rs32,~Rt32) + C Intrinsic Prototype: Word32 Q6_R_andor_RnR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andor_RnR __builtin_HEXAGON_M4_or_andn + +/* ========================================================================== + Assembly Syntax: Rx32|=or(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_oror_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_oror_RR __builtin_HEXAGON_M4_or_or + +/* ========================================================================== + Assembly Syntax: Rx32|=xor(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_xoror_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_xoror_RR __builtin_HEXAGON_M4_or_xor + +/* ========================================================================== + Assembly Syntax: Rdd32=pmpyw(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_pmpyw_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_pmpyw_RR __builtin_HEXAGON_M4_pmpyw + +/* ========================================================================== + Assembly Syntax: Rxx32^=pmpyw(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_pmpywxacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_pmpywxacc_RR __builtin_HEXAGON_M4_pmpyw_acc + +/* ========================================================================== + Assembly Syntax: Rdd32=vpmpyh(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vpmpyh_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vpmpyh_RR __builtin_HEXAGON_M4_vpmpyh + +/* ========================================================================== + Assembly Syntax: Rxx32^=vpmpyh(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vpmpyhxacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vpmpyhxacc_RR __builtin_HEXAGON_M4_vpmpyh_acc + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpyweh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpywehacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpywehacc_PP __builtin_HEXAGON_M4_vrmpyeh_acc_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpyweh(Rss32,Rtt32):<<1 + C Intrinsic Prototype: Word64 Q6_P_vrmpywehacc_PP_s1(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpywehacc_PP_s1 __builtin_HEXAGON_M4_vrmpyeh_acc_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpyweh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpyweh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpyweh_PP __builtin_HEXAGON_M4_vrmpyeh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpyweh(Rss32,Rtt32):<<1 + C Intrinsic Prototype: Word64 Q6_P_vrmpyweh_PP_s1(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpyweh_PP_s1 __builtin_HEXAGON_M4_vrmpyeh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpywoh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpywohacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpywohacc_PP __builtin_HEXAGON_M4_vrmpyoh_acc_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpywoh(Rss32,Rtt32):<<1 + C Intrinsic Prototype: Word64 Q6_P_vrmpywohacc_PP_s1(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpywohacc_PP_s1 __builtin_HEXAGON_M4_vrmpyoh_acc_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpywoh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpywoh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpywoh_PP __builtin_HEXAGON_M4_vrmpyoh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpywoh(Rss32,Rtt32):<<1 + C Intrinsic Prototype: Word64 Q6_P_vrmpywoh_PP_s1(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpywoh_PP_s1 __builtin_HEXAGON_M4_vrmpyoh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32^=and(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_andxacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andxacc_RR __builtin_HEXAGON_M4_xor_and + +/* ========================================================================== + Assembly Syntax: Rx32^=and(Rs32,~Rt32) + C Intrinsic Prototype: Word32 Q6_R_andxacc_RnR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andxacc_RnR __builtin_HEXAGON_M4_xor_andn + +/* ========================================================================== + Assembly Syntax: Rx32^=or(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_orxacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_orxacc_RR __builtin_HEXAGON_M4_xor_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=xor(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_xorxacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_xorxacc_PP __builtin_HEXAGON_M4_xor_xacc + +/* ========================================================================== + Assembly Syntax: Rxx32+=vdmpybsu(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vdmpybsuacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vdmpybsuacc_PP_sat __builtin_HEXAGON_M5_vdmacbsu + +/* ========================================================================== + Assembly Syntax: Rdd32=vdmpybsu(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vdmpybsu_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vdmpybsu_PP_sat __builtin_HEXAGON_M5_vdmpybsu + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpybsu(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vmpybsuacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpybsuacc_RR __builtin_HEXAGON_M5_vmacbsu + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpybu(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vmpybuacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpybuacc_RR __builtin_HEXAGON_M5_vmacbuu + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpybsu(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vmpybsu_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpybsu_RR __builtin_HEXAGON_M5_vmpybsu + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpybu(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vmpybu_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpybu_RR __builtin_HEXAGON_M5_vmpybuu + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpybsu(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpybsuacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpybsuacc_PP __builtin_HEXAGON_M5_vrmacbsu + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpybu(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpybuacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpybuacc_PP __builtin_HEXAGON_M5_vrmacbuu + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpybsu(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpybsu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpybsu_PP __builtin_HEXAGON_M5_vrmpybsu + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpybu(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpybu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpybu_PP __builtin_HEXAGON_M5_vrmpybuu + +/* ========================================================================== + Assembly Syntax: Rd32=addasl(Rt32,Rs32,#u3) + C Intrinsic Prototype: Word32 Q6_R_addasl_RRI(Word32 Rt, Word32 Rs, Word32 Iu3) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_addasl_RRI __builtin_HEXAGON_S2_addasl_rrri + +/* ========================================================================== + Assembly Syntax: Rdd32=asl(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asl_PI(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asl_PI __builtin_HEXAGON_S2_asl_i_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=asl(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_aslacc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslacc_PI __builtin_HEXAGON_S2_asl_i_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=asl(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asland_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asland_PI __builtin_HEXAGON_S2_asl_i_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=asl(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_aslnac_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslnac_PI __builtin_HEXAGON_S2_asl_i_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=asl(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_aslor_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslor_PI __builtin_HEXAGON_S2_asl_i_p_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=asl(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_aslxacc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslxacc_PI __builtin_HEXAGON_S2_asl_i_p_xacc + +/* ========================================================================== + Assembly Syntax: Rd32=asl(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asl_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asl_RI __builtin_HEXAGON_S2_asl_i_r + +/* ========================================================================== + Assembly Syntax: Rx32+=asl(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_aslacc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslacc_RI __builtin_HEXAGON_S2_asl_i_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=asl(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asland_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asland_RI __builtin_HEXAGON_S2_asl_i_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=asl(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_aslnac_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslnac_RI __builtin_HEXAGON_S2_asl_i_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=asl(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_aslor_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslor_RI __builtin_HEXAGON_S2_asl_i_r_or + +/* ========================================================================== + Assembly Syntax: Rd32=asl(Rs32,#u5):sat + C Intrinsic Prototype: Word32 Q6_R_asl_RI_sat(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asl_RI_sat __builtin_HEXAGON_S2_asl_i_r_sat + +/* ========================================================================== + Assembly Syntax: Rx32^=asl(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_aslxacc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslxacc_RI __builtin_HEXAGON_S2_asl_i_r_xacc + +/* ========================================================================== + Assembly Syntax: Rdd32=vaslh(Rss32,#u4) + C Intrinsic Prototype: Word64 Q6_P_vaslh_PI(Word64 Rss, Word32 Iu4) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaslh_PI __builtin_HEXAGON_S2_asl_i_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vaslw(Rss32,#u5) + C Intrinsic Prototype: Word64 Q6_P_vaslw_PI(Word64 Rss, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaslw_PI __builtin_HEXAGON_S2_asl_i_vw + +/* ========================================================================== + Assembly Syntax: Rdd32=asl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asl_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asl_PR __builtin_HEXAGON_S2_asl_r_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=asl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_aslacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslacc_PR __builtin_HEXAGON_S2_asl_r_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=asl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asland_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asland_PR __builtin_HEXAGON_S2_asl_r_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=asl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_aslnac_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslnac_PR __builtin_HEXAGON_S2_asl_r_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=asl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_aslor_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslor_PR __builtin_HEXAGON_S2_asl_r_p_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=asl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_aslxacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslxacc_PR __builtin_HEXAGON_S2_asl_r_p_xor + +/* ========================================================================== + Assembly Syntax: Rd32=asl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asl_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asl_RR __builtin_HEXAGON_S2_asl_r_r + +/* ========================================================================== + Assembly Syntax: Rx32+=asl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_aslacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslacc_RR __builtin_HEXAGON_S2_asl_r_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=asl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asland_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asland_RR __builtin_HEXAGON_S2_asl_r_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=asl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_aslnac_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslnac_RR __builtin_HEXAGON_S2_asl_r_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=asl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_aslor_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslor_RR __builtin_HEXAGON_S2_asl_r_r_or + +/* ========================================================================== + Assembly Syntax: Rd32=asl(Rs32,Rt32):sat + C Intrinsic Prototype: Word32 Q6_R_asl_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asl_RR_sat __builtin_HEXAGON_S2_asl_r_r_sat + +/* ========================================================================== + Assembly Syntax: Rdd32=vaslh(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vaslh_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaslh_PR __builtin_HEXAGON_S2_asl_r_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vaslw(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vaslw_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaslw_PR __builtin_HEXAGON_S2_asl_r_vw + +/* ========================================================================== + Assembly Syntax: Rdd32=asr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asr_PI(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asr_PI __builtin_HEXAGON_S2_asr_i_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=asr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asracc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asracc_PI __builtin_HEXAGON_S2_asr_i_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=asr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asrand_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asrand_PI __builtin_HEXAGON_S2_asr_i_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=asr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asrnac_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asrnac_PI __builtin_HEXAGON_S2_asr_i_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=asr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asror_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asror_PI __builtin_HEXAGON_S2_asr_i_p_or + +/* ========================================================================== + Assembly Syntax: Rdd32=asr(Rss32,#u6):rnd + C Intrinsic Prototype: Word64 Q6_P_asr_PI_rnd(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asr_PI_rnd __builtin_HEXAGON_S2_asr_i_p_rnd + +/* ========================================================================== + Assembly Syntax: Rdd32=asrrnd(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asrrnd_PI(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_asrrnd_PI __builtin_HEXAGON_S2_asr_i_p_rnd_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rd32=asr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asr_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asr_RI __builtin_HEXAGON_S2_asr_i_r + +/* ========================================================================== + Assembly Syntax: Rx32+=asr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asracc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asracc_RI __builtin_HEXAGON_S2_asr_i_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=asr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asrand_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asrand_RI __builtin_HEXAGON_S2_asr_i_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=asr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asrnac_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asrnac_RI __builtin_HEXAGON_S2_asr_i_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=asr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asror_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asror_RI __builtin_HEXAGON_S2_asr_i_r_or + +/* ========================================================================== + Assembly Syntax: Rd32=asr(Rs32,#u5):rnd + C Intrinsic Prototype: Word32 Q6_R_asr_RI_rnd(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asr_RI_rnd __builtin_HEXAGON_S2_asr_i_r_rnd + +/* ========================================================================== + Assembly Syntax: Rd32=asrrnd(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asrrnd_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_asrrnd_RI __builtin_HEXAGON_S2_asr_i_r_rnd_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rd32=vasrw(Rss32,#u5) + C Intrinsic Prototype: Word32 Q6_R_vasrw_PI(Word64 Rss, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vasrw_PI __builtin_HEXAGON_S2_asr_i_svw_trun + +/* ========================================================================== + Assembly Syntax: Rdd32=vasrh(Rss32,#u4) + C Intrinsic Prototype: Word64 Q6_P_vasrh_PI(Word64 Rss, Word32 Iu4) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vasrh_PI __builtin_HEXAGON_S2_asr_i_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vasrw(Rss32,#u5) + C Intrinsic Prototype: Word64 Q6_P_vasrw_PI(Word64 Rss, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vasrw_PI __builtin_HEXAGON_S2_asr_i_vw + +/* ========================================================================== + Assembly Syntax: Rdd32=asr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asr_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asr_PR __builtin_HEXAGON_S2_asr_r_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=asr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asracc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asracc_PR __builtin_HEXAGON_S2_asr_r_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=asr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asrand_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asrand_PR __builtin_HEXAGON_S2_asr_r_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=asr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asrnac_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asrnac_PR __builtin_HEXAGON_S2_asr_r_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=asr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asror_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asror_PR __builtin_HEXAGON_S2_asr_r_p_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=asr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asrxacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asrxacc_PR __builtin_HEXAGON_S2_asr_r_p_xor + +/* ========================================================================== + Assembly Syntax: Rd32=asr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asr_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asr_RR __builtin_HEXAGON_S2_asr_r_r + +/* ========================================================================== + Assembly Syntax: Rx32+=asr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asracc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asracc_RR __builtin_HEXAGON_S2_asr_r_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=asr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asrand_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asrand_RR __builtin_HEXAGON_S2_asr_r_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=asr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asrnac_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asrnac_RR __builtin_HEXAGON_S2_asr_r_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=asr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asror_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asror_RR __builtin_HEXAGON_S2_asr_r_r_or + +/* ========================================================================== + Assembly Syntax: Rd32=asr(Rs32,Rt32):sat + C Intrinsic Prototype: Word32 Q6_R_asr_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asr_RR_sat __builtin_HEXAGON_S2_asr_r_r_sat + +/* ========================================================================== + Assembly Syntax: Rd32=vasrw(Rss32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_vasrw_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vasrw_PR __builtin_HEXAGON_S2_asr_r_svw_trun + +/* ========================================================================== + Assembly Syntax: Rdd32=vasrh(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vasrh_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vasrh_PR __builtin_HEXAGON_S2_asr_r_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vasrw(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vasrw_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vasrw_PR __builtin_HEXAGON_S2_asr_r_vw + +/* ========================================================================== + Assembly Syntax: Rd32=brev(Rs32) + C Intrinsic Prototype: Word32 Q6_R_brev_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_brev_R __builtin_HEXAGON_S2_brev + +/* ========================================================================== + Assembly Syntax: Rdd32=brev(Rss32) + C Intrinsic Prototype: Word64 Q6_P_brev_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_brev_P __builtin_HEXAGON_S2_brevp + +/* ========================================================================== + Assembly Syntax: Rd32=cl0(Rs32) + C Intrinsic Prototype: Word32 Q6_R_cl0_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cl0_R __builtin_HEXAGON_S2_cl0 + +/* ========================================================================== + Assembly Syntax: Rd32=cl0(Rss32) + C Intrinsic Prototype: Word32 Q6_R_cl0_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cl0_P __builtin_HEXAGON_S2_cl0p + +/* ========================================================================== + Assembly Syntax: Rd32=cl1(Rs32) + C Intrinsic Prototype: Word32 Q6_R_cl1_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cl1_R __builtin_HEXAGON_S2_cl1 + +/* ========================================================================== + Assembly Syntax: Rd32=cl1(Rss32) + C Intrinsic Prototype: Word32 Q6_R_cl1_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cl1_P __builtin_HEXAGON_S2_cl1p + +/* ========================================================================== + Assembly Syntax: Rd32=clb(Rs32) + C Intrinsic Prototype: Word32 Q6_R_clb_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_clb_R __builtin_HEXAGON_S2_clb + +/* ========================================================================== + Assembly Syntax: Rd32=normamt(Rs32) + C Intrinsic Prototype: Word32 Q6_R_normamt_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_normamt_R __builtin_HEXAGON_S2_clbnorm + +/* ========================================================================== + Assembly Syntax: Rd32=clb(Rss32) + C Intrinsic Prototype: Word32 Q6_R_clb_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_clb_P __builtin_HEXAGON_S2_clbp + +/* ========================================================================== + Assembly Syntax: Rd32=clrbit(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_clrbit_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_clrbit_RI __builtin_HEXAGON_S2_clrbit_i + +/* ========================================================================== + Assembly Syntax: Rd32=clrbit(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_clrbit_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_clrbit_RR __builtin_HEXAGON_S2_clrbit_r + +/* ========================================================================== + Assembly Syntax: Rd32=ct0(Rs32) + C Intrinsic Prototype: Word32 Q6_R_ct0_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_ct0_R __builtin_HEXAGON_S2_ct0 + +/* ========================================================================== + Assembly Syntax: Rd32=ct0(Rss32) + C Intrinsic Prototype: Word32 Q6_R_ct0_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_ct0_P __builtin_HEXAGON_S2_ct0p + +/* ========================================================================== + Assembly Syntax: Rd32=ct1(Rs32) + C Intrinsic Prototype: Word32 Q6_R_ct1_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_ct1_R __builtin_HEXAGON_S2_ct1 + +/* ========================================================================== + Assembly Syntax: Rd32=ct1(Rss32) + C Intrinsic Prototype: Word32 Q6_R_ct1_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_ct1_P __builtin_HEXAGON_S2_ct1p + +/* ========================================================================== + Assembly Syntax: Rdd32=deinterleave(Rss32) + C Intrinsic Prototype: Word64 Q6_P_deinterleave_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_deinterleave_P __builtin_HEXAGON_S2_deinterleave + +/* ========================================================================== + Assembly Syntax: Rd32=extractu(Rs32,#u5,#U5) + C Intrinsic Prototype: Word32 Q6_R_extractu_RII(Word32 Rs, Word32 Iu5, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_extractu_RII __builtin_HEXAGON_S2_extractu + +/* ========================================================================== + Assembly Syntax: Rd32=extractu(Rs32,Rtt32) + C Intrinsic Prototype: Word32 Q6_R_extractu_RP(Word32 Rs, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_extractu_RP __builtin_HEXAGON_S2_extractu_rp + +/* ========================================================================== + Assembly Syntax: Rdd32=extractu(Rss32,#u6,#U6) + C Intrinsic Prototype: Word64 Q6_P_extractu_PII(Word64 Rss, Word32 Iu6, Word32 IU6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_extractu_PII __builtin_HEXAGON_S2_extractup + +/* ========================================================================== + Assembly Syntax: Rdd32=extractu(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_extractu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_extractu_PP __builtin_HEXAGON_S2_extractup_rp + +/* ========================================================================== + Assembly Syntax: Rx32=insert(Rs32,#u5,#U5) + C Intrinsic Prototype: Word32 Q6_R_insert_RII(Word32 Rx, Word32 Rs, Word32 Iu5, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_insert_RII __builtin_HEXAGON_S2_insert + +/* ========================================================================== + Assembly Syntax: Rx32=insert(Rs32,Rtt32) + C Intrinsic Prototype: Word32 Q6_R_insert_RP(Word32 Rx, Word32 Rs, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_insert_RP __builtin_HEXAGON_S2_insert_rp + +/* ========================================================================== + Assembly Syntax: Rxx32=insert(Rss32,#u6,#U6) + C Intrinsic Prototype: Word64 Q6_P_insert_PII(Word64 Rxx, Word64 Rss, Word32 Iu6, Word32 IU6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_insert_PII __builtin_HEXAGON_S2_insertp + +/* ========================================================================== + Assembly Syntax: Rxx32=insert(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_insert_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_insert_PP __builtin_HEXAGON_S2_insertp_rp + +/* ========================================================================== + Assembly Syntax: Rdd32=interleave(Rss32) + C Intrinsic Prototype: Word64 Q6_P_interleave_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_interleave_P __builtin_HEXAGON_S2_interleave + +/* ========================================================================== + Assembly Syntax: Rdd32=lfs(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_lfs_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lfs_PP __builtin_HEXAGON_S2_lfsp + +/* ========================================================================== + Assembly Syntax: Rdd32=lsl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsl_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsl_PR __builtin_HEXAGON_S2_lsl_r_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=lsl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lslacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lslacc_PR __builtin_HEXAGON_S2_lsl_r_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=lsl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsland_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsland_PR __builtin_HEXAGON_S2_lsl_r_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=lsl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lslnac_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lslnac_PR __builtin_HEXAGON_S2_lsl_r_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=lsl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lslor_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lslor_PR __builtin_HEXAGON_S2_lsl_r_p_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=lsl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lslxacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lslxacc_PR __builtin_HEXAGON_S2_lsl_r_p_xor + +/* ========================================================================== + Assembly Syntax: Rd32=lsl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsl_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsl_RR __builtin_HEXAGON_S2_lsl_r_r + +/* ========================================================================== + Assembly Syntax: Rx32+=lsl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lslacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lslacc_RR __builtin_HEXAGON_S2_lsl_r_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=lsl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsland_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsland_RR __builtin_HEXAGON_S2_lsl_r_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=lsl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lslnac_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lslnac_RR __builtin_HEXAGON_S2_lsl_r_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=lsl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lslor_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lslor_RR __builtin_HEXAGON_S2_lsl_r_r_or + +/* ========================================================================== + Assembly Syntax: Rdd32=vlslh(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vlslh_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vlslh_PR __builtin_HEXAGON_S2_lsl_r_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vlslw(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vlslw_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vlslw_PR __builtin_HEXAGON_S2_lsl_r_vw + +/* ========================================================================== + Assembly Syntax: Rdd32=lsr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_lsr_PI(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsr_PI __builtin_HEXAGON_S2_lsr_i_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=lsr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_lsracc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsracc_PI __builtin_HEXAGON_S2_lsr_i_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=lsr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_lsrand_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsrand_PI __builtin_HEXAGON_S2_lsr_i_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=lsr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_lsrnac_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsrnac_PI __builtin_HEXAGON_S2_lsr_i_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=lsr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_lsror_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsror_PI __builtin_HEXAGON_S2_lsr_i_p_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=lsr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_lsrxacc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsrxacc_PI __builtin_HEXAGON_S2_lsr_i_p_xacc + +/* ========================================================================== + Assembly Syntax: Rd32=lsr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_lsr_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsr_RI __builtin_HEXAGON_S2_lsr_i_r + +/* ========================================================================== + Assembly Syntax: Rx32+=lsr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_lsracc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsracc_RI __builtin_HEXAGON_S2_lsr_i_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=lsr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_lsrand_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsrand_RI __builtin_HEXAGON_S2_lsr_i_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=lsr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_lsrnac_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsrnac_RI __builtin_HEXAGON_S2_lsr_i_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=lsr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_lsror_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsror_RI __builtin_HEXAGON_S2_lsr_i_r_or + +/* ========================================================================== + Assembly Syntax: Rx32^=lsr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_lsrxacc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsrxacc_RI __builtin_HEXAGON_S2_lsr_i_r_xacc + +/* ========================================================================== + Assembly Syntax: Rdd32=vlsrh(Rss32,#u4) + C Intrinsic Prototype: Word64 Q6_P_vlsrh_PI(Word64 Rss, Word32 Iu4) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vlsrh_PI __builtin_HEXAGON_S2_lsr_i_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vlsrw(Rss32,#u5) + C Intrinsic Prototype: Word64 Q6_P_vlsrw_PI(Word64 Rss, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vlsrw_PI __builtin_HEXAGON_S2_lsr_i_vw + +/* ========================================================================== + Assembly Syntax: Rdd32=lsr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsr_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsr_PR __builtin_HEXAGON_S2_lsr_r_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=lsr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsracc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsracc_PR __builtin_HEXAGON_S2_lsr_r_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=lsr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsrand_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsrand_PR __builtin_HEXAGON_S2_lsr_r_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=lsr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsrnac_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsrnac_PR __builtin_HEXAGON_S2_lsr_r_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=lsr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsror_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsror_PR __builtin_HEXAGON_S2_lsr_r_p_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=lsr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsrxacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsrxacc_PR __builtin_HEXAGON_S2_lsr_r_p_xor + +/* ========================================================================== + Assembly Syntax: Rd32=lsr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsr_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsr_RR __builtin_HEXAGON_S2_lsr_r_r + +/* ========================================================================== + Assembly Syntax: Rx32+=lsr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsracc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsracc_RR __builtin_HEXAGON_S2_lsr_r_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=lsr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsrand_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsrand_RR __builtin_HEXAGON_S2_lsr_r_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=lsr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsrnac_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsrnac_RR __builtin_HEXAGON_S2_lsr_r_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=lsr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsror_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsror_RR __builtin_HEXAGON_S2_lsr_r_r_or + +/* ========================================================================== + Assembly Syntax: Rdd32=vlsrh(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vlsrh_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vlsrh_PR __builtin_HEXAGON_S2_lsr_r_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vlsrw(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vlsrw_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vlsrw_PR __builtin_HEXAGON_S2_lsr_r_vw + +/* ========================================================================== + Assembly Syntax: Rdd32=packhl(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_packhl_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_packhl_RR __builtin_HEXAGON_S2_packhl + +/* ========================================================================== + Assembly Syntax: Rd32=parity(Rss32,Rtt32) + C Intrinsic Prototype: Word32 Q6_R_parity_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_parity_PP __builtin_HEXAGON_S2_parityp + +/* ========================================================================== + Assembly Syntax: Rd32=setbit(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_setbit_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_setbit_RI __builtin_HEXAGON_S2_setbit_i + +/* ========================================================================== + Assembly Syntax: Rd32=setbit(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_setbit_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_setbit_RR __builtin_HEXAGON_S2_setbit_r + +/* ========================================================================== + Assembly Syntax: Rdd32=shuffeb(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_shuffeb_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_shuffeb_PP __builtin_HEXAGON_S2_shuffeb + +/* ========================================================================== + Assembly Syntax: Rdd32=shuffeh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_shuffeh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_shuffeh_PP __builtin_HEXAGON_S2_shuffeh + +/* ========================================================================== + Assembly Syntax: Rdd32=shuffob(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_shuffob_PP(Word64 Rtt, Word64 Rss) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_shuffob_PP __builtin_HEXAGON_S2_shuffob + +/* ========================================================================== + Assembly Syntax: Rdd32=shuffoh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_shuffoh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_shuffoh_PP __builtin_HEXAGON_S2_shuffoh + +/* ========================================================================== + Assembly Syntax: memb(Rx32++#s4:0:circ(Mu2))=Rt32 + C Intrinsic Prototype: void Q6_memb_IMR_circ(void** Rx, Word32 Is4_0, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memb_IMR_circ __builtin_HEXAGON_S2_storerb_pci + +/* ========================================================================== + Assembly Syntax: memb(Rx32++I:circ(Mu2))=Rt32 + C Intrinsic Prototype: void Q6_memb_MR_circ(void** Rx, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memb_MR_circ __builtin_HEXAGON_S2_storerb_pcr + +/* ========================================================================== + Assembly Syntax: memd(Rx32++#s4:3:circ(Mu2))=Rtt32 + C Intrinsic Prototype: void Q6_memd_IMP_circ(void** Rx, Word32 Is4_3, Word32 Mu, Word64 Rtt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memd_IMP_circ __builtin_HEXAGON_S2_storerd_pci + +/* ========================================================================== + Assembly Syntax: memd(Rx32++I:circ(Mu2))=Rtt32 + C Intrinsic Prototype: void Q6_memd_MP_circ(void** Rx, Word32 Mu, Word64 Rtt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memd_MP_circ __builtin_HEXAGON_S2_storerd_pcr + +/* ========================================================================== + Assembly Syntax: memh(Rx32++#s4:1:circ(Mu2))=Rt32.h + C Intrinsic Prototype: void Q6_memh_IMRh_circ(void** Rx, Word32 Is4_1, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memh_IMRh_circ __builtin_HEXAGON_S2_storerf_pci + +/* ========================================================================== + Assembly Syntax: memh(Rx32++I:circ(Mu2))=Rt32.h + C Intrinsic Prototype: void Q6_memh_MRh_circ(void** Rx, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memh_MRh_circ __builtin_HEXAGON_S2_storerf_pcr + +/* ========================================================================== + Assembly Syntax: memh(Rx32++#s4:1:circ(Mu2))=Rt32 + C Intrinsic Prototype: void Q6_memh_IMR_circ(void** Rx, Word32 Is4_1, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memh_IMR_circ __builtin_HEXAGON_S2_storerh_pci + +/* ========================================================================== + Assembly Syntax: memh(Rx32++I:circ(Mu2))=Rt32 + C Intrinsic Prototype: void Q6_memh_MR_circ(void** Rx, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memh_MR_circ __builtin_HEXAGON_S2_storerh_pcr + +/* ========================================================================== + Assembly Syntax: memw(Rx32++#s4:2:circ(Mu2))=Rt32 + C Intrinsic Prototype: void Q6_memw_IMR_circ(void** Rx, Word32 Is4_2, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memw_IMR_circ __builtin_HEXAGON_S2_storeri_pci + +/* ========================================================================== + Assembly Syntax: memw(Rx32++I:circ(Mu2))=Rt32 + C Intrinsic Prototype: void Q6_memw_MR_circ(void** Rx, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memw_MR_circ __builtin_HEXAGON_S2_storeri_pcr + +/* ========================================================================== + Assembly Syntax: Rd32=vsathb(Rs32) + C Intrinsic Prototype: Word32 Q6_R_vsathb_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsathb_R __builtin_HEXAGON_S2_svsathb + +/* ========================================================================== + Assembly Syntax: Rd32=vsathub(Rs32) + C Intrinsic Prototype: Word32 Q6_R_vsathub_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsathub_R __builtin_HEXAGON_S2_svsathub + +/* ========================================================================== + Assembly Syntax: Rx32=tableidxb(Rs32,#u4,#U5) + C Intrinsic Prototype: Word32 Q6_R_tableidxb_RII(Word32 Rx, Word32 Rs, Word32 Iu4, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_tableidxb_RII __builtin_HEXAGON_S2_tableidxb_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rx32=tableidxd(Rs32,#u4,#U5) + C Intrinsic Prototype: Word32 Q6_R_tableidxd_RII(Word32 Rx, Word32 Rs, Word32 Iu4, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_tableidxd_RII __builtin_HEXAGON_S2_tableidxd_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rx32=tableidxh(Rs32,#u4,#U5) + C Intrinsic Prototype: Word32 Q6_R_tableidxh_RII(Word32 Rx, Word32 Rs, Word32 Iu4, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_tableidxh_RII __builtin_HEXAGON_S2_tableidxh_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rx32=tableidxw(Rs32,#u4,#U5) + C Intrinsic Prototype: Word32 Q6_R_tableidxw_RII(Word32 Rx, Word32 Rs, Word32 Iu4, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_tableidxw_RII __builtin_HEXAGON_S2_tableidxw_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rd32=togglebit(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_togglebit_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_togglebit_RI __builtin_HEXAGON_S2_togglebit_i + +/* ========================================================================== + Assembly Syntax: Rd32=togglebit(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_togglebit_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_togglebit_RR __builtin_HEXAGON_S2_togglebit_r + +/* ========================================================================== + Assembly Syntax: Pd4=tstbit(Rs32,#u5) + C Intrinsic Prototype: Byte Q6_p_tstbit_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_tstbit_RI __builtin_HEXAGON_S2_tstbit_i + +/* ========================================================================== + Assembly Syntax: Pd4=tstbit(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_tstbit_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_tstbit_RR __builtin_HEXAGON_S2_tstbit_r + +/* ========================================================================== + Assembly Syntax: Rdd32=valignb(Rtt32,Rss32,#u3) + C Intrinsic Prototype: Word64 Q6_P_valignb_PPI(Word64 Rtt, Word64 Rss, Word32 Iu3) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_valignb_PPI __builtin_HEXAGON_S2_valignib + +/* ========================================================================== + Assembly Syntax: Rdd32=valignb(Rtt32,Rss32,Pu4) + C Intrinsic Prototype: Word64 Q6_P_valignb_PPp(Word64 Rtt, Word64 Rss, Byte Pu) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_valignb_PPp __builtin_HEXAGON_S2_valignrb + +/* ========================================================================== + Assembly Syntax: Rdd32=vcnegh(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vcnegh_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcnegh_PR __builtin_HEXAGON_S2_vcnegh + +/* ========================================================================== + Assembly Syntax: Rdd32=vcrotate(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vcrotate_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcrotate_PR __builtin_HEXAGON_S2_vcrotate + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcnegh(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vrcneghacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcneghacc_PR __builtin_HEXAGON_S2_vrcnegh + +/* ========================================================================== + Assembly Syntax: Rd32=vrndwh(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vrndwh_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vrndwh_P __builtin_HEXAGON_S2_vrndpackwh + +/* ========================================================================== + Assembly Syntax: Rd32=vrndwh(Rss32):sat + C Intrinsic Prototype: Word32 Q6_R_vrndwh_P_sat(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vrndwh_P_sat __builtin_HEXAGON_S2_vrndpackwhs + +/* ========================================================================== + Assembly Syntax: Rd32=vsathb(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vsathb_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsathb_P __builtin_HEXAGON_S2_vsathb + +/* ========================================================================== + Assembly Syntax: Rdd32=vsathb(Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsathb_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsathb_P __builtin_HEXAGON_S2_vsathb_nopack + +/* ========================================================================== + Assembly Syntax: Rd32=vsathub(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vsathub_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsathub_P __builtin_HEXAGON_S2_vsathub + +/* ========================================================================== + Assembly Syntax: Rdd32=vsathub(Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsathub_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsathub_P __builtin_HEXAGON_S2_vsathub_nopack + +/* ========================================================================== + Assembly Syntax: Rd32=vsatwh(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vsatwh_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsatwh_P __builtin_HEXAGON_S2_vsatwh + +/* ========================================================================== + Assembly Syntax: Rdd32=vsatwh(Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsatwh_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsatwh_P __builtin_HEXAGON_S2_vsatwh_nopack + +/* ========================================================================== + Assembly Syntax: Rd32=vsatwuh(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vsatwuh_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsatwuh_P __builtin_HEXAGON_S2_vsatwuh + +/* ========================================================================== + Assembly Syntax: Rdd32=vsatwuh(Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsatwuh_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsatwuh_P __builtin_HEXAGON_S2_vsatwuh_nopack + +/* ========================================================================== + Assembly Syntax: Rd32=vsplatb(Rs32) + C Intrinsic Prototype: Word32 Q6_R_vsplatb_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsplatb_R __builtin_HEXAGON_S2_vsplatrb + +/* ========================================================================== + Assembly Syntax: Rdd32=vsplath(Rs32) + C Intrinsic Prototype: Word64 Q6_P_vsplath_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsplath_R __builtin_HEXAGON_S2_vsplatrh + +/* ========================================================================== + Assembly Syntax: Rdd32=vspliceb(Rss32,Rtt32,#u3) + C Intrinsic Prototype: Word64 Q6_P_vspliceb_PPI(Word64 Rss, Word64 Rtt, Word32 Iu3) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vspliceb_PPI __builtin_HEXAGON_S2_vspliceib + +/* ========================================================================== + Assembly Syntax: Rdd32=vspliceb(Rss32,Rtt32,Pu4) + C Intrinsic Prototype: Word64 Q6_P_vspliceb_PPp(Word64 Rss, Word64 Rtt, Byte Pu) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vspliceb_PPp __builtin_HEXAGON_S2_vsplicerb + +/* ========================================================================== + Assembly Syntax: Rdd32=vsxtbh(Rs32) + C Intrinsic Prototype: Word64 Q6_P_vsxtbh_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsxtbh_R __builtin_HEXAGON_S2_vsxtbh + +/* ========================================================================== + Assembly Syntax: Rdd32=vsxthw(Rs32) + C Intrinsic Prototype: Word64 Q6_P_vsxthw_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsxthw_R __builtin_HEXAGON_S2_vsxthw + +/* ========================================================================== + Assembly Syntax: Rd32=vtrunehb(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vtrunehb_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vtrunehb_P __builtin_HEXAGON_S2_vtrunehb + +/* ========================================================================== + Assembly Syntax: Rdd32=vtrunewh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vtrunewh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vtrunewh_PP __builtin_HEXAGON_S2_vtrunewh + +/* ========================================================================== + Assembly Syntax: Rd32=vtrunohb(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vtrunohb_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vtrunohb_P __builtin_HEXAGON_S2_vtrunohb + +/* ========================================================================== + Assembly Syntax: Rdd32=vtrunowh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vtrunowh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vtrunowh_PP __builtin_HEXAGON_S2_vtrunowh + +/* ========================================================================== + Assembly Syntax: Rdd32=vzxtbh(Rs32) + C Intrinsic Prototype: Word64 Q6_P_vzxtbh_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vzxtbh_R __builtin_HEXAGON_S2_vzxtbh + +/* ========================================================================== + Assembly Syntax: Rdd32=vzxthw(Rs32) + C Intrinsic Prototype: Word64 Q6_P_vzxthw_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vzxthw_R __builtin_HEXAGON_S2_vzxthw + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rs32,add(Ru32,#s6)) + C Intrinsic Prototype: Word32 Q6_R_add_add_RRI(Word32 Rs, Word32 Ru, Word32 Is6) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_add_RRI __builtin_HEXAGON_S4_addaddi + +/* ========================================================================== + Assembly Syntax: Rx32=add(#u8,asl(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_add_asl_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_asl_IRI __builtin_HEXAGON_S4_addi_asl_ri + +/* ========================================================================== + Assembly Syntax: Rx32=add(#u8,lsr(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_add_lsr_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_lsr_IRI __builtin_HEXAGON_S4_addi_lsr_ri + +/* ========================================================================== + Assembly Syntax: Rx32=and(#u8,asl(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_and_asl_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_and_asl_IRI __builtin_HEXAGON_S4_andi_asl_ri + +/* ========================================================================== + Assembly Syntax: Rx32=and(#u8,lsr(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_and_lsr_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_and_lsr_IRI __builtin_HEXAGON_S4_andi_lsr_ri + +/* ========================================================================== + Assembly Syntax: Rd32=add(clb(Rs32),#s6) + C Intrinsic Prototype: Word32 Q6_R_add_clb_RI(Word32 Rs, Word32 Is6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_clb_RI __builtin_HEXAGON_S4_clbaddi + +/* ========================================================================== + Assembly Syntax: Rd32=add(clb(Rss32),#s6) + C Intrinsic Prototype: Word32 Q6_R_add_clb_PI(Word64 Rss, Word32 Is6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_clb_PI __builtin_HEXAGON_S4_clbpaddi + +/* ========================================================================== + Assembly Syntax: Rd32=normamt(Rss32) + C Intrinsic Prototype: Word32 Q6_R_normamt_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_normamt_P __builtin_HEXAGON_S4_clbpnorm + +/* ========================================================================== + Assembly Syntax: Rd32=extract(Rs32,#u5,#U5) + C Intrinsic Prototype: Word32 Q6_R_extract_RII(Word32 Rs, Word32 Iu5, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_extract_RII __builtin_HEXAGON_S4_extract + +/* ========================================================================== + Assembly Syntax: Rd32=extract(Rs32,Rtt32) + C Intrinsic Prototype: Word32 Q6_R_extract_RP(Word32 Rs, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_extract_RP __builtin_HEXAGON_S4_extract_rp + +/* ========================================================================== + Assembly Syntax: Rdd32=extract(Rss32,#u6,#U6) + C Intrinsic Prototype: Word64 Q6_P_extract_PII(Word64 Rss, Word32 Iu6, Word32 IU6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_extract_PII __builtin_HEXAGON_S4_extractp + +/* ========================================================================== + Assembly Syntax: Rdd32=extract(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_extract_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_extract_PP __builtin_HEXAGON_S4_extractp_rp + +/* ========================================================================== + Assembly Syntax: Rd32=lsl(#s6,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsl_IR(Word32 Is6, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsl_IR __builtin_HEXAGON_S4_lsli + +/* ========================================================================== + Assembly Syntax: Pd4=!tstbit(Rs32,#u5) + C Intrinsic Prototype: Byte Q6_p_not_tstbit_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_tstbit_RI __builtin_HEXAGON_S4_ntstbit_i + +/* ========================================================================== + Assembly Syntax: Pd4=!tstbit(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_not_tstbit_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_tstbit_RR __builtin_HEXAGON_S4_ntstbit_r + +/* ========================================================================== + Assembly Syntax: Rx32|=and(Rs32,#s10) + C Intrinsic Prototype: Word32 Q6_R_andor_RI(Word32 Rx, Word32 Rs, Word32 Is10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andor_RI __builtin_HEXAGON_S4_or_andi + +/* ========================================================================== + Assembly Syntax: Rx32=or(Ru32,and(Rx32,#s10)) + C Intrinsic Prototype: Word32 Q6_R_or_and_RRI(Word32 Ru, Word32 Rx, Word32 Is10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_or_and_RRI __builtin_HEXAGON_S4_or_andix + +/* ========================================================================== + Assembly Syntax: Rx32|=or(Rs32,#s10) + C Intrinsic Prototype: Word32 Q6_R_oror_RI(Word32 Rx, Word32 Rs, Word32 Is10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_oror_RI __builtin_HEXAGON_S4_or_ori + +/* ========================================================================== + Assembly Syntax: Rx32=or(#u8,asl(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_or_asl_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_or_asl_IRI __builtin_HEXAGON_S4_ori_asl_ri + +/* ========================================================================== + Assembly Syntax: Rx32=or(#u8,lsr(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_or_lsr_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_or_lsr_IRI __builtin_HEXAGON_S4_ori_lsr_ri + +/* ========================================================================== + Assembly Syntax: Rd32=parity(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_parity_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_parity_RR __builtin_HEXAGON_S4_parity + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rs32,sub(#s6,Ru32)) + C Intrinsic Prototype: Word32 Q6_R_add_sub_RIR(Word32 Rs, Word32 Is6, Word32 Ru) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_sub_RIR __builtin_HEXAGON_S4_subaddi + +/* ========================================================================== + Assembly Syntax: Rx32=sub(#u8,asl(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_sub_asl_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_asl_IRI __builtin_HEXAGON_S4_subi_asl_ri + +/* ========================================================================== + Assembly Syntax: Rx32=sub(#u8,lsr(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_sub_lsr_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_lsr_IRI __builtin_HEXAGON_S4_subi_lsr_ri + +/* ========================================================================== + Assembly Syntax: Rdd32=vrcrotate(Rss32,Rt32,#u2) + C Intrinsic Prototype: Word64 Q6_P_vrcrotate_PRI(Word64 Rss, Word32 Rt, Word32 Iu2) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcrotate_PRI __builtin_HEXAGON_S4_vrcrotate + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcrotate(Rss32,Rt32,#u2) + C Intrinsic Prototype: Word64 Q6_P_vrcrotateacc_PRI(Word64 Rxx, Word64 Rss, Word32 Rt, Word32 Iu2) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcrotateacc_PRI __builtin_HEXAGON_S4_vrcrotate_acc + +/* ========================================================================== + Assembly Syntax: Rdd32=vxaddsubh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vxaddsubh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vxaddsubh_PP_sat __builtin_HEXAGON_S4_vxaddsubh + +/* ========================================================================== + Assembly Syntax: Rdd32=vxaddsubh(Rss32,Rtt32):rnd:>>1:sat + C Intrinsic Prototype: Word64 Q6_P_vxaddsubh_PP_rnd_rs1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vxaddsubh_PP_rnd_rs1_sat __builtin_HEXAGON_S4_vxaddsubhr + +/* ========================================================================== + Assembly Syntax: Rdd32=vxaddsubw(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vxaddsubw_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vxaddsubw_PP_sat __builtin_HEXAGON_S4_vxaddsubw + +/* ========================================================================== + Assembly Syntax: Rdd32=vxsubaddh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vxsubaddh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vxsubaddh_PP_sat __builtin_HEXAGON_S4_vxsubaddh + +/* ========================================================================== + Assembly Syntax: Rdd32=vxsubaddh(Rss32,Rtt32):rnd:>>1:sat + C Intrinsic Prototype: Word64 Q6_P_vxsubaddh_PP_rnd_rs1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vxsubaddh_PP_rnd_rs1_sat __builtin_HEXAGON_S4_vxsubaddhr + +/* ========================================================================== + Assembly Syntax: Rdd32=vxsubaddw(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vxsubaddw_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vxsubaddw_PP_sat __builtin_HEXAGON_S4_vxsubaddw + +/* ========================================================================== + Assembly Syntax: Rd32=vasrhub(Rss32,#u4):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_vasrhub_PI_rnd_sat(Word64 Rss, Word32 Iu4) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vasrhub_PI_rnd_sat __builtin_HEXAGON_S5_asrhub_rnd_sat_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rd32=vasrhub(Rss32,#u4):sat + C Intrinsic Prototype: Word32 Q6_R_vasrhub_PI_sat(Word64 Rss, Word32 Iu4) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vasrhub_PI_sat __builtin_HEXAGON_S5_asrhub_sat + +/* ========================================================================== + Assembly Syntax: Rd32=popcount(Rss32) + C Intrinsic Prototype: Word32 Q6_R_popcount_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_popcount_P __builtin_HEXAGON_S5_popcountp + +/* ========================================================================== + Assembly Syntax: Rdd32=vasrh(Rss32,#u4):rnd + C Intrinsic Prototype: Word64 Q6_P_vasrh_PI_rnd(Word64 Rss, Word32 Iu4) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_vasrh_PI_rnd __builtin_HEXAGON_S5_vasrhrnd_goodsyntax + +/* ========================================================================== + Assembly Syntax: dccleana(Rs32) + C Intrinsic Prototype: void Q6_dccleana_A(Address Rs) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dccleana_A __builtin_HEXAGON_Y2_dccleana + +/* ========================================================================== + Assembly Syntax: dccleaninva(Rs32) + C Intrinsic Prototype: void Q6_dccleaninva_A(Address Rs) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dccleaninva_A __builtin_HEXAGON_Y2_dccleaninva + +/* ========================================================================== + Assembly Syntax: dcfetch(Rs32) + C Intrinsic Prototype: void Q6_dcfetch_A(Address Rs) + Instruction Type: MAPPING + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_dcfetch_A __builtin_HEXAGON_Y2_dcfetch + +/* ========================================================================== + Assembly Syntax: dcinva(Rs32) + C Intrinsic Prototype: void Q6_dcinva_A(Address Rs) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dcinva_A __builtin_HEXAGON_Y2_dcinva + +/* ========================================================================== + Assembly Syntax: dczeroa(Rs32) + C Intrinsic Prototype: void Q6_dczeroa_A(Address Rs) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dczeroa_A __builtin_HEXAGON_Y2_dczeroa + +/* ========================================================================== + Assembly Syntax: l2fetch(Rs32,Rt32) + C Intrinsic Prototype: void Q6_l2fetch_AR(Address Rs, Word32 Rt) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_l2fetch_AR __builtin_HEXAGON_Y4_l2fetch + +/* ========================================================================== + Assembly Syntax: l2fetch(Rs32,Rtt32) + C Intrinsic Prototype: void Q6_l2fetch_AP(Address Rs, Word64 Rtt) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_l2fetch_AP __builtin_HEXAGON_Y5_l2fetch + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rdd32=rol(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_rol_PI(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_rol_PI __builtin_HEXAGON_S6_rol_i_p +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rxx32+=rol(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_rolacc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_rolacc_PI __builtin_HEXAGON_S6_rol_i_p_acc +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rxx32&=rol(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_roland_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_roland_PI __builtin_HEXAGON_S6_rol_i_p_and +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rxx32-=rol(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_rolnac_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_rolnac_PI __builtin_HEXAGON_S6_rol_i_p_nac +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rxx32|=rol(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_rolor_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_rolor_PI __builtin_HEXAGON_S6_rol_i_p_or +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rxx32^=rol(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_rolxacc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_rolxacc_PI __builtin_HEXAGON_S6_rol_i_p_xacc +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rd32=rol(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_rol_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_rol_RI __builtin_HEXAGON_S6_rol_i_r +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rx32+=rol(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_rolacc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_rolacc_RI __builtin_HEXAGON_S6_rol_i_r_acc +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rx32&=rol(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_roland_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_roland_RI __builtin_HEXAGON_S6_rol_i_r_and +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rx32-=rol(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_rolnac_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_rolnac_RI __builtin_HEXAGON_S6_rol_i_r_nac +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rx32|=rol(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_rolor_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_rolor_RI __builtin_HEXAGON_S6_rol_i_r_or +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rx32^=rol(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_rolxacc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_rolxacc_RI __builtin_HEXAGON_S6_rol_i_r_xacc +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Rdd32=vabsdiffb(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vabsdiffb_PP(Word64 Rtt, Word64 Rss) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsdiffb_PP __builtin_HEXAGON_M6_vabsdiffb +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HEXAGON_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Rdd32=vabsdiffub(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vabsdiffub_PP(Word64 Rtt, Word64 Rss) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsdiffub_PP __builtin_HEXAGON_M6_vabsdiffub +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HEXAGON_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Rdd32=vsplatb(Rs32) + C Intrinsic Prototype: Word64 Q6_P_vsplatb_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsplatb_R __builtin_HEXAGON_S6_vsplatrbp +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HEXAGON_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Rdd32=vtrunehb(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vtrunehb_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vtrunehb_PP __builtin_HEXAGON_S6_vtrunehb_ppp +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HEXAGON_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Rdd32=vtrunohb(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vtrunohb_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vtrunohb_PP __builtin_HEXAGON_S6_vtrunohb_ppp +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HEXAGON_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32=vmem(Rt32):nt + C Intrinsic Prototype: HVX_Vector Q6_V_vmem_R_nt(Word32 Rt) + Instruction Type: MAPPING + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vmem_R_nt __builtin_HEXAGON_V6_ldntnt0 +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HEXAGON_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Pd4=!any8(vcmpb.eq(Rss32,Rtt32)) + C Intrinsic Prototype: Byte Q6_p_not_any8_vcmpb_eq_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_any8_vcmpb_eq_PP __builtin_HEXAGON_A6_vcmpbeq_notany +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HEXAGON_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Rdd32=dfadd(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfadd_PP(Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfadd_PP __builtin_HEXAGON_F2_dfadd +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HEXAGON_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Rdd32=dfsub(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfsub_PP(Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfsub_PP __builtin_HEXAGON_F2_dfsub +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HEXAGON_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Rx32-=mpyi(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mpyinac_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyinac_RR __builtin_HEXAGON_M2_mnaci +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HEXAGON_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Rd32=mask(#u5,#U5) + C Intrinsic Prototype: Word32 Q6_R_mask_II(Word32 Iu5, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mask_II __builtin_HEXAGON_S2_mask +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=clip(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_clip_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_clip_RI __builtin_HEXAGON_A7_clip +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=cround(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_cround_PI(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cround_PI __builtin_HEXAGON_A7_croundd_ri +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=cround(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_cround_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cround_PR __builtin_HEXAGON_A7_croundd_rr +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=vclip(Rss32,#u5) + C Intrinsic Prototype: Word64 Q6_P_vclip_PI(Word64 Rss, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vclip_PI __builtin_HEXAGON_A7_vclip +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 +/* ========================================================================== + Assembly Syntax: Rdd32=dfmax(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfmax_PP(Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmax_PP __builtin_HEXAGON_F2_dfmax +#endif /* __HEXAGON_ARCH___ >= 67 */ + +#if __HEXAGON_ARCH__ >= 67 +/* ========================================================================== + Assembly Syntax: Rdd32=dfmin(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfmin_PP(Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmin_PP __builtin_HEXAGON_F2_dfmin +#endif /* __HEXAGON_ARCH___ >= 67 */ + +#if __HEXAGON_ARCH__ >= 67 +/* ========================================================================== + Assembly Syntax: Rdd32=dfmpyfix(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfmpyfix_PP(Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmpyfix_PP __builtin_HEXAGON_F2_dfmpyfix +#endif /* __HEXAGON_ARCH___ >= 67 */ + +#if __HEXAGON_ARCH__ >= 67 +/* ========================================================================== + Assembly Syntax: Rxx32+=dfmpyhh(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfmpyhhacc_PP(Float64 Rxx, Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmpyhhacc_PP __builtin_HEXAGON_F2_dfmpyhh +#endif /* __HEXAGON_ARCH___ >= 67 */ + +#if __HEXAGON_ARCH__ >= 67 +/* ========================================================================== + Assembly Syntax: Rxx32+=dfmpylh(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfmpylhacc_PP(Float64 Rxx, Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmpylhacc_PP __builtin_HEXAGON_F2_dfmpylh +#endif /* __HEXAGON_ARCH___ >= 67 */ + +#if __HEXAGON_ARCH__ >= 67 +/* ========================================================================== + Assembly Syntax: Rdd32=dfmpyll(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfmpyll_PP(Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmpyll_PP __builtin_HEXAGON_F2_dfmpyll +#endif /* __HEXAGON_ARCH___ >= 67 */ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=cmpyiw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyiw_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyiw_PP __builtin_HEXAGON_M7_dcmpyiw +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpyiw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyiwacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyiwacc_PP __builtin_HEXAGON_M7_dcmpyiw_acc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=cmpyiw(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_cmpyiw_PP_conj(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyiw_PP_conj __builtin_HEXAGON_M7_dcmpyiwc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpyiw(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_cmpyiwacc_PP_conj(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyiwacc_PP_conj __builtin_HEXAGON_M7_dcmpyiwc_acc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=cmpyrw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyrw_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyrw_PP __builtin_HEXAGON_M7_dcmpyrw +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpyrw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyrwacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyrwacc_PP __builtin_HEXAGON_M7_dcmpyrw_acc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=cmpyrw(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_cmpyrw_PP_conj(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyrw_PP_conj __builtin_HEXAGON_M7_dcmpyrwc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpyrw(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_cmpyrwacc_PP_conj(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyrwacc_PP_conj __builtin_HEXAGON_M7_dcmpyrwc_acc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=vdmpyw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vdmpyw_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_vdmpyw_PP __builtin_HEXAGON_M7_vdmpy +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rxx32+=vdmpyw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vdmpywacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_vdmpywacc_PP __builtin_HEXAGON_M7_vdmpy_acc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyiw(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyiw_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyiw_PP_s1_sat __builtin_HEXAGON_M7_wcmpyiw +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyiw(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyiw_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyiw_PP_s1_rnd_sat __builtin_HEXAGON_M7_wcmpyiw_rnd +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyiw(Rss32,Rtt32*):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyiw_PP_conj_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyiw_PP_conj_s1_sat __builtin_HEXAGON_M7_wcmpyiwc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyiw(Rss32,Rtt32*):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyiw_PP_conj_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyiw_PP_conj_s1_rnd_sat __builtin_HEXAGON_M7_wcmpyiwc_rnd +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyrw(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyrw_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyrw_PP_s1_sat __builtin_HEXAGON_M7_wcmpyrw +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyrw(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyrw_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyrw_PP_s1_rnd_sat __builtin_HEXAGON_M7_wcmpyrw_rnd +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyrw(Rss32,Rtt32*):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyrw_PP_conj_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyrw_PP_conj_s1_sat __builtin_HEXAGON_M7_wcmpyrwc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyrw(Rss32,Rtt32*):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyrw_PP_conj_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyrw_PP_conj_s1_rnd_sat __builtin_HEXAGON_M7_wcmpyrwc_rnd +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: dmlink(Rs32,Rt32) + C Intrinsic Prototype: void Q6_dmlink_AA(Address Rs, Address Rt) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dmlink_AA __builtin_HEXAGON_Y6_dmlink +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HEXAGON_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Rd32=dmpause + C Intrinsic Prototype: Word32 Q6_R_dmpause() + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_R_dmpause __builtin_HEXAGON_Y6_dmpause +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HEXAGON_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Rd32=dmpoll + C Intrinsic Prototype: Word32 Q6_R_dmpoll() + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_R_dmpoll __builtin_HEXAGON_Y6_dmpoll +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HEXAGON_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: dmresume(Rs32) + C Intrinsic Prototype: void Q6_dmresume_A(Address Rs) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dmresume_A __builtin_HEXAGON_Y6_dmresume +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HEXAGON_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: dmstart(Rs32) + C Intrinsic Prototype: void Q6_dmstart_A(Address Rs) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dmstart_A __builtin_HEXAGON_Y6_dmstart +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HEXAGON_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Rd32=dmwait + C Intrinsic Prototype: Word32 Q6_R_dmwait() + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_R_dmwait __builtin_HEXAGON_Y6_dmwait +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#include +#ifdef __HVX__ +#include +#endif /* __HVX__ */ +#endif diff --git a/lib/include/hexagon_types.h b/lib/include/hexagon_types.h new file mode 100644 index 0000000000..6958809418 --- /dev/null +++ b/lib/include/hexagon_types.h @@ -0,0 +1,2653 @@ +/******************************************************************************/ +/* (c) 2020 Qualcomm Innovation Center, Inc. All rights reserved. */ +/* */ +/******************************************************************************/ +#ifndef HEXAGON_TYPES_H +#define HEXAGON_TYPES_H + +#include + +/* Hexagon names */ +#define HEXAGON_Vect HEXAGON_Vect64 +#define HEXAGON_V_GET_D HEXAGON_V64_GET_D +#define HEXAGON_V_GET_UD HEXAGON_V64_GET_UD +#define HEXAGON_V_GET_W0 HEXAGON_V64_GET_W0 +#define HEXAGON_V_GET_W1 HEXAGON_V64_GET_W1 +#define HEXAGON_V_GET_UW0 HEXAGON_V64_GET_UW0 +#define HEXAGON_V_GET_UW1 HEXAGON_V64_GET_UW1 +#define HEXAGON_V_GET_H0 HEXAGON_V64_GET_H0 +#define HEXAGON_V_GET_H1 HEXAGON_V64_GET_H1 +#define HEXAGON_V_GET_H2 HEXAGON_V64_GET_H2 +#define HEXAGON_V_GET_H3 HEXAGON_V64_GET_H3 +#define HEXAGON_V_GET_UH0 HEXAGON_V64_GET_UH0 +#define HEXAGON_V_GET_UH1 HEXAGON_V64_GET_UH1 +#define HEXAGON_V_GET_UH2 HEXAGON_V64_GET_UH2 +#define HEXAGON_V_GET_UH3 HEXAGON_V64_GET_UH3 +#define HEXAGON_V_GET_B0 HEXAGON_V64_GET_B0 +#define HEXAGON_V_GET_B1 HEXAGON_V64_GET_B1 +#define HEXAGON_V_GET_B2 HEXAGON_V64_GET_B2 +#define HEXAGON_V_GET_B3 HEXAGON_V64_GET_B3 +#define HEXAGON_V_GET_B4 HEXAGON_V64_GET_B4 +#define HEXAGON_V_GET_B5 HEXAGON_V64_GET_B5 +#define HEXAGON_V_GET_B6 HEXAGON_V64_GET_B6 +#define HEXAGON_V_GET_B7 HEXAGON_V64_GET_B7 +#define HEXAGON_V_GET_UB0 HEXAGON_V64_GET_UB0 +#define HEXAGON_V_GET_UB1 HEXAGON_V64_GET_UB1 +#define HEXAGON_V_GET_UB2 HEXAGON_V64_GET_UB2 +#define HEXAGON_V_GET_UB3 HEXAGON_V64_GET_UB3 +#define HEXAGON_V_GET_UB4 HEXAGON_V64_GET_UB4 +#define HEXAGON_V_GET_UB5 HEXAGON_V64_GET_UB5 +#define HEXAGON_V_GET_UB6 HEXAGON_V64_GET_UB6 +#define HEXAGON_V_GET_UB7 HEXAGON_V64_GET_UB7 +#define HEXAGON_V_PUT_D HEXAGON_V64_PUT_D +#define HEXAGON_V_PUT_W0 HEXAGON_V64_PUT_W0 +#define HEXAGON_V_PUT_W1 HEXAGON_V64_PUT_W1 +#define HEXAGON_V_PUT_H0 HEXAGON_V64_PUT_H0 +#define HEXAGON_V_PUT_H1 HEXAGON_V64_PUT_H1 +#define HEXAGON_V_PUT_H2 HEXAGON_V64_PUT_H2 +#define HEXAGON_V_PUT_H3 HEXAGON_V64_PUT_H3 +#define HEXAGON_V_PUT_B0 HEXAGON_V64_PUT_B0 +#define HEXAGON_V_PUT_B1 HEXAGON_V64_PUT_B1 +#define HEXAGON_V_PUT_B2 HEXAGON_V64_PUT_B2 +#define HEXAGON_V_PUT_B3 HEXAGON_V64_PUT_B3 +#define HEXAGON_V_PUT_B4 HEXAGON_V64_PUT_B4 +#define HEXAGON_V_PUT_B5 HEXAGON_V64_PUT_B5 +#define HEXAGON_V_PUT_B6 HEXAGON_V64_PUT_B6 +#define HEXAGON_V_PUT_B7 HEXAGON_V64_PUT_B7 +#define HEXAGON_V_CREATE_D HEXAGON_V64_CREATE_D +#define HEXAGON_V_CREATE_W HEXAGON_V64_CREATE_W +#define HEXAGON_V_CREATE_H HEXAGON_V64_CREATE_H +#define HEXAGON_V_CREATE_B HEXAGON_V64_CREATE_B + +#ifdef __cplusplus +#define HEXAGON_VectC HEXAGON_Vect64C +#endif /* __cplusplus */ + +/* 64 Bit Vectors */ + +typedef long long __attribute__((__may_alias__)) HEXAGON_Vect64; + +/* Extract doubleword macros */ + +#define HEXAGON_V64_GET_D(v) (v) +#define HEXAGON_V64_GET_UD(v) ((unsigned long long)(v)) + +/* Extract word macros */ + +#define HEXAGON_V64_GET_W0(v) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.w[0]; \ + }) +#define HEXAGON_V64_GET_W1(v) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.w[1]; \ + }) +#define HEXAGON_V64_GET_UW0(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned int uw[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.uw[0]; \ + }) +#define HEXAGON_V64_GET_UW1(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned int uw[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.uw[1]; \ + }) + +/* Extract half word macros */ + +#define HEXAGON_V64_GET_H0(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[0]; \ + }) +#define HEXAGON_V64_GET_H1(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[1]; \ + }) +#define HEXAGON_V64_GET_H2(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[2]; \ + }) +#define HEXAGON_V64_GET_H3(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[3]; \ + }) +#define HEXAGON_V64_GET_UH0(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.uh[0]; \ + }) +#define HEXAGON_V64_GET_UH1(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.uh[1]; \ + }) +#define HEXAGON_V64_GET_UH2(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.uh[2]; \ + }) +#define HEXAGON_V64_GET_UH3(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.uh[3]; \ + }) + +/* Extract byte macros */ + +#define HEXAGON_V64_GET_B0(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[0]; \ + }) +#define HEXAGON_V64_GET_B1(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[1]; \ + }) +#define HEXAGON_V64_GET_B2(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[2]; \ + }) +#define HEXAGON_V64_GET_B3(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[3]; \ + }) +#define HEXAGON_V64_GET_B4(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[4]; \ + }) +#define HEXAGON_V64_GET_B5(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[5]; \ + }) +#define HEXAGON_V64_GET_B6(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[6]; \ + }) +#define HEXAGON_V64_GET_B7(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[7]; \ + }) +#define HEXAGON_V64_GET_UB0(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[0]; \ + }) +#define HEXAGON_V64_GET_UB1(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[1]; \ + }) +#define HEXAGON_V64_GET_UB2(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[2]; \ + }) +#define HEXAGON_V64_GET_UB3(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[3]; \ + }) +#define HEXAGON_V64_GET_UB4(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[4]; \ + }) +#define HEXAGON_V64_GET_UB5(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[5]; \ + }) +#define HEXAGON_V64_GET_UB6(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[6]; \ + }) +#define HEXAGON_V64_GET_UB7(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[7]; \ + }) + +/* NOTE: All set macros return a HEXAGON_Vect64 type */ + +/* Set doubleword macro */ + +#define HEXAGON_V64_PUT_D(v, new) (new) + +/* Set word macros */ + +#ifdef __hexagon__ + +#define HEXAGON_V64_PUT_W0(v, new) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.w[0] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_W1(v, new) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.w[1] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V64_PUT_W0(v, new) \ + (((v) & 0xffffffff00000000LL) | ((HEXAGON_Vect64)((unsigned int)(new)))) +#define HEXAGON_V64_PUT_W1(v, new) \ + (((v) & 0x00000000ffffffffLL) | (((HEXAGON_Vect64)(new)) << 32LL)) + +#endif /* !__hexagon__ */ + +/* Set half word macros */ + +#ifdef __hexagon__ + +#define HEXAGON_V64_PUT_H0(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[0] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_H1(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[1] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_H2(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[2] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_H3(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[3] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V64_PUT_H0(v, new) \ + (((v) & 0xffffffffffff0000LL) | ((HEXAGON_Vect64)((unsigned short)(new)))) +#define HEXAGON_V64_PUT_H1(v, new) \ + (((v) & 0xffffffff0000ffffLL) | (((HEXAGON_Vect64)((unsigned short)(new))) << 16LL)) +#define HEXAGON_V64_PUT_H2(v, new) \ + (((v) & 0xffff0000ffffffffLL) | (((HEXAGON_Vect64)((unsigned short)(new))) << 32LL)) +#define HEXAGON_V64_PUT_H3(v, new) \ + (((v) & 0x0000ffffffffffffLL) | (((HEXAGON_Vect64)(new)) << 48LL)) + +#endif /* !__hexagon__ */ + +/* Set byte macros */ + +#ifdef __hexagon__ + +#define HEXAGON_V64_PUT_B0(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[0] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B1(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[1] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B2(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[2] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B3(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[3] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B4(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[4] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B5(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[5] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B6(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[6] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B7(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[7] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V64_PUT_B0(v, new) \ + (((v) & 0xffffffffffffff00LL) | ((HEXAGON_Vect64)((unsigned char)(new)))) +#define HEXAGON_V64_PUT_B1(v, new) \ + (((v) & 0xffffffffffff00ffLL) | (((HEXAGON_Vect64)((unsigned char)(new))) << 8LL)) +#define HEXAGON_V64_PUT_B2(v, new) \ + (((v) & 0xffffffffff00ffffLL) | (((HEXAGON_Vect64)((unsigned char)(new))) << 16LL)) +#define HEXAGON_V64_PUT_B3(v, new) \ + (((v) & 0xffffffff00ffffffLL) | (((HEXAGON_Vect64)((unsigned char)(new))) << 24LL)) +#define HEXAGON_V64_PUT_B4(v, new) \ + (((v) & 0xffffff00ffffffffLL) | (((HEXAGON_Vect64)((unsigned char)(new))) << 32LL)) +#define HEXAGON_V64_PUT_B5(v, new) \ + (((v) & 0xffff00ffffffffffLL) | (((HEXAGON_Vect64)((unsigned char)(new))) << 40LL)) +#define HEXAGON_V64_PUT_B6(v, new) \ + (((v) & 0xff00ffffffffffffLL) | (((HEXAGON_Vect64)((unsigned char)(new))) << 48LL)) +#define HEXAGON_V64_PUT_B7(v, new) \ + (((v) & 0x00ffffffffffffffLL) | (((HEXAGON_Vect64)(new)) << 56LL)) + +#endif /* !__hexagon__ */ + +/* NOTE: All create macros return a HEXAGON_Vect64 type */ + +/* Create from a doubleword */ + +#define HEXAGON_V64_CREATE_D(d) (d) + +/* Create from words */ + +#ifdef __hexagon__ + +#define HEXAGON_V64_CREATE_W(w1, w0) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.w[0] = (w0); \ + _HEXAGON_V64_internal_union.w[1] = (w1); \ + _HEXAGON_V64_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V64_CREATE_W(w1, w0) \ + ((((HEXAGON_Vect64)(w1)) << 32LL) | ((HEXAGON_Vect64)((w0) & 0xffffffff))) + +#endif /* !__hexagon__ */ + +/* Create from half words */ + +#ifdef __hexagon__ + +#define HEXAGON_V64_CREATE_H(h3, h2, h1, h0) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.h[0] = (h0); \ + _HEXAGON_V64_internal_union.h[1] = (h1); \ + _HEXAGON_V64_internal_union.h[2] = (h2); \ + _HEXAGON_V64_internal_union.h[3] = (h3); \ + _HEXAGON_V64_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V64_CREATE_H(h3, h2, h1, h0) \ + ((((HEXAGON_Vect64)(h3)) << 48LL) | (((HEXAGON_Vect64)((h2) & 0xffff)) << 32LL) | \ + (((HEXAGON_Vect64)((h1) & 0xffff)) << 16LL) | ((HEXAGON_Vect64)((h0) & 0xffff))) + +#endif /* !__hexagon__ */ + +/* Create from bytes */ + +#ifdef __hexagon__ + +#define HEXAGON_V64_CREATE_B(b7, b6, b5, b4, b3, b2, b1, b0) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.b[0] = (b0); \ + _HEXAGON_V64_internal_union.b[1] = (b1); \ + _HEXAGON_V64_internal_union.b[2] = (b2); \ + _HEXAGON_V64_internal_union.b[3] = (b3); \ + _HEXAGON_V64_internal_union.b[4] = (b4); \ + _HEXAGON_V64_internal_union.b[5] = (b5); \ + _HEXAGON_V64_internal_union.b[6] = (b6); \ + _HEXAGON_V64_internal_union.b[7] = (b7); \ + _HEXAGON_V64_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V64_CREATE_B(b7, b6, b5, b4, b3, b2, b1, b0) \ + ((((HEXAGON_Vect64)(b7)) << 56LL) | (((HEXAGON_Vect64)((b6) & 0xff)) << 48LL) | \ + (((HEXAGON_Vect64)((b5) & 0xff)) << 40LL) | (((HEXAGON_Vect64)((b4) & 0xff)) << 32LL) | \ + (((HEXAGON_Vect64)((b3) & 0xff)) << 24LL) | (((HEXAGON_Vect64)((b2) & 0xff)) << 16LL) | \ + (((HEXAGON_Vect64)((b1) & 0xff)) << 8LL) | ((HEXAGON_Vect64)((b0) & 0xff))) + +#endif /* !__hexagon__ */ + +#ifdef __cplusplus + +class HEXAGON_Vect64C { +public: + // Constructors + HEXAGON_Vect64C(long long d = 0) : data(d) {}; + HEXAGON_Vect64C(int w1, int w0) : data(HEXAGON_V64_CREATE_W(w1, w0)) {}; + HEXAGON_Vect64C(short h3, short h2, short h1, short h0) + : data(HEXAGON_V64_CREATE_H(h3, h2, h1, h0)) {}; + HEXAGON_Vect64C(signed char b7, signed char b6, signed char b5, signed char b4, + signed char b3, signed char b2, signed char b1, signed char b0) + : data(HEXAGON_V64_CREATE_B(b7, b6, b5, b4, b3, b2, b1, b0)) {}; + HEXAGON_Vect64C(const HEXAGON_Vect64C &v) : data(v.data) {}; + + HEXAGON_Vect64C &operator=(const HEXAGON_Vect64C &v) { + data = v.data; + return *this; + }; + + operator long long() { + return data; + }; + + // Extract doubleword methods + long long D(void) { + return HEXAGON_V64_GET_D(data); + }; + unsigned long long UD(void) { + return HEXAGON_V64_GET_UD(data); + }; + + // Extract word methods + int W0(void) { + return HEXAGON_V64_GET_W0(data); + }; + int W1(void) { + return HEXAGON_V64_GET_W1(data); + }; + unsigned int UW0(void) { + return HEXAGON_V64_GET_UW0(data); + }; + unsigned int UW1(void) { + return HEXAGON_V64_GET_UW1(data); + }; + + // Extract half word methods + short H0(void) { + return HEXAGON_V64_GET_H0(data); + }; + short H1(void) { + return HEXAGON_V64_GET_H1(data); + }; + short H2(void) { + return HEXAGON_V64_GET_H2(data); + }; + short H3(void) { + return HEXAGON_V64_GET_H3(data); + }; + unsigned short UH0(void) { + return HEXAGON_V64_GET_UH0(data); + }; + unsigned short UH1(void) { + return HEXAGON_V64_GET_UH1(data); + }; + unsigned short UH2(void) { + return HEXAGON_V64_GET_UH2(data); + }; + unsigned short UH3(void) { + return HEXAGON_V64_GET_UH3(data); + }; + + // Extract byte methods + signed char B0(void) { + return HEXAGON_V64_GET_B0(data); + }; + signed char B1(void) { + return HEXAGON_V64_GET_B1(data); + }; + signed char B2(void) { + return HEXAGON_V64_GET_B2(data); + }; + signed char B3(void) { + return HEXAGON_V64_GET_B3(data); + }; + signed char B4(void) { + return HEXAGON_V64_GET_B4(data); + }; + signed char B5(void) { + return HEXAGON_V64_GET_B5(data); + }; + signed char B6(void) { + return HEXAGON_V64_GET_B6(data); + }; + signed char B7(void) { + return HEXAGON_V64_GET_B7(data); + }; + unsigned char UB0(void) { + return HEXAGON_V64_GET_UB0(data); + }; + unsigned char UB1(void) { + return HEXAGON_V64_GET_UB1(data); + }; + unsigned char UB2(void) { + return HEXAGON_V64_GET_UB2(data); + }; + unsigned char UB3(void) { + return HEXAGON_V64_GET_UB3(data); + }; + unsigned char UB4(void) { + return HEXAGON_V64_GET_UB4(data); + }; + unsigned char UB5(void) { + return HEXAGON_V64_GET_UB5(data); + }; + unsigned char UB6(void) { + return HEXAGON_V64_GET_UB6(data); + }; + unsigned char UB7(void) { + return HEXAGON_V64_GET_UB7(data); + }; + + // NOTE: All set methods return a HEXAGON_Vect64C type + + // Set doubleword method + HEXAGON_Vect64C D(long long d) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_D(data, d)); + }; + + // Set word methods + HEXAGON_Vect64C W0(int w) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_W0(data, w)); + }; + HEXAGON_Vect64C W1(int w) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_W1(data, w)); + }; + + // Set half word methods + HEXAGON_Vect64C H0(short h) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_H0(data, h)); + }; + HEXAGON_Vect64C H1(short h) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_H1(data, h)); + }; + HEXAGON_Vect64C H2(short h) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_H2(data, h)); + }; + HEXAGON_Vect64C H3(short h) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_H3(data, h)); + }; + + // Set byte methods + HEXAGON_Vect64C B0(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B0(data, b)); + }; + HEXAGON_Vect64C B1(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B1(data, b)); + }; + HEXAGON_Vect64C B2(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B2(data, b)); + }; + HEXAGON_Vect64C B3(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B3(data, b)); + }; + HEXAGON_Vect64C B4(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B4(data, b)); + }; + HEXAGON_Vect64C B5(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B5(data, b)); + }; + HEXAGON_Vect64C B6(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B6(data, b)); + }; + HEXAGON_Vect64C B7(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B7(data, b)); + }; + +private: + long long data; +}; + +#endif /* __cplusplus */ + +/* 32 Bit Vectors */ + +typedef int HEXAGON_Vect32; + +/* Extract word macros */ + +#define HEXAGON_V32_GET_W(v) (v) +#define HEXAGON_V32_GET_UW(v) ((unsigned int)(v)) + +/* Extract half word macros */ + +#define HEXAGON_V32_GET_H0(v) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.h[0]; \ + }) +#define HEXAGON_V32_GET_H1(v) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.h[1]; \ + }) +#define HEXAGON_V32_GET_UH0(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned short uh[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.uh[0]; \ + }) +#define HEXAGON_V32_GET_UH1(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned short uh[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.uh[1]; \ + }) + +/* Extract byte macros */ + +#define HEXAGON_V32_GET_B0(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[0]; \ + }) +#define HEXAGON_V32_GET_B1(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[1]; \ + }) +#define HEXAGON_V32_GET_B2(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[2]; \ + }) +#define HEXAGON_V32_GET_B3(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[3]; \ + }) +#define HEXAGON_V32_GET_UB0(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.ub[0]; \ + }) +#define HEXAGON_V32_GET_UB1(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.ub[1]; \ + }) +#define HEXAGON_V32_GET_UB2(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.ub[2]; \ + }) +#define HEXAGON_V32_GET_UB3(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.ub[3]; \ + }) + +/* NOTE: All set macros return a HEXAGON_Vect32 type */ + +/* Set word macro */ + +#define HEXAGON_V32_PUT_W(v, new) (new) + +/* Set half word macros */ + +#ifdef __hexagon__ + +#define HEXAGON_V32_PUT_H0(v, new) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.h[0] = (new); \ + _HEXAGON_V32_internal_union.w; \ + }) +#define HEXAGON_V32_PUT_H1(v, new) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.h[1] = (new); \ + _HEXAGON_V32_internal_union.w; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V32_PUT_H0(v, new) \ + (((v) & 0xffff0000) | ((HEXAGON_Vect32)((unsigned short)(new)))) +#define HEXAGON_V32_PUT_H1(v, new) (((v) & 0x0000ffff) | (((HEXAGON_Vect32)(new)) << 16)) + +#endif /* !__hexagon__ */ + +/* Set byte macros */ + +#ifdef __hexagon__ + +#define HEXAGON_V32_PUT_B0(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[0] = (new); \ + _HEXAGON_V32_internal_union.w; \ + }) +#define HEXAGON_V32_PUT_B1(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[1] = (new); \ + _HEXAGON_V32_internal_union.w; \ + }) +#define HEXAGON_V32_PUT_B2(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[2] = (new); \ + _HEXAGON_V32_internal_union.w; \ + }) +#define HEXAGON_V32_PUT_B3(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[3] = (new); \ + _HEXAGON_V32_internal_union.w; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V32_PUT_B0(v, new) \ + (((v) & 0xffffff00) | ((HEXAGON_Vect32)((unsigned char)(new)))) +#define HEXAGON_V32_PUT_B1(v, new) \ + (((v) & 0xffff00ff) | (((HEXAGON_Vect32)((unsigned char)(new))) << 8)) +#define HEXAGON_V32_PUT_B2(v, new) \ + (((v) & 0xff00ffff) | (((HEXAGON_Vect32)((unsigned char)(new))) << 16)) +#define HEXAGON_V32_PUT_B3(v, new) (((v) & 0x00ffffff) | (((HEXAGON_Vect32)(new)) << 24)) + +#endif /* !__hexagon__ */ + +/* NOTE: All create macros return a HEXAGON_Vect32 type */ + +/* Create from a word */ + +#define HEXAGON_V32_CREATE_W(w) (w) + +/* Create from half words */ + +#ifdef __hexagon__ + +#define HEXAGON_V32_CREATE_H(h1, h0) \ + __extension__({ \ + union { \ + long long d; \ + short h[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.h[0] = (h0); \ + _HEXAGON_V32_internal_union.h[1] = (h1); \ + _HEXAGON_V32_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V32_CREATE_H(h1, h0) \ + ((((HEXAGON_Vect32)(h1)) << 16) | ((HEXAGON_Vect32)((h0) & 0xffff))) + +#endif /* !__hexagon__ */ + +/* Create from bytes */ +#ifdef __hexagon__ + +#define HEXAGON_V32_CREATE_B(b3, b2, b1, b0) \ + __extension__({ \ + union { \ + long long d; \ + char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.b[0] = (b0); \ + _HEXAGON_V32_internal_union.b[1] = (b1); \ + _HEXAGON_V32_internal_union.b[2] = (b2); \ + _HEXAGON_V32_internal_union.b[3] = (b3); \ + _HEXAGON_V32_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V32_CREATE_B(b3, b2, b1, b0) \ + ((((HEXAGON_Vect32)(b3)) << 24) | (((HEXAGON_Vect32)((b2) & 0xff)) << 16) | \ + (((HEXAGON_Vect32)((b1) & 0xff)) << 8) | ((HEXAGON_Vect32)((b0) & 0xff))) + +#endif /* !__hexagon__ */ + +#ifdef __cplusplus + +class HEXAGON_Vect32C { +public: + // Constructors + HEXAGON_Vect32C(int w = 0) : data(w) {}; + HEXAGON_Vect32C(short h1, short h0) : data(HEXAGON_V32_CREATE_H(h1, h0)) {}; + HEXAGON_Vect32C(signed char b3, signed char b2, signed char b1, signed char b0) + : data(HEXAGON_V32_CREATE_B(b3, b2, b1, b0)) {}; + HEXAGON_Vect32C(const HEXAGON_Vect32C &v) : data(v.data) {}; + + HEXAGON_Vect32C &operator=(const HEXAGON_Vect32C &v) { + data = v.data; + return *this; + }; + + operator int() { + return data; + }; + + // Extract word methods + int W(void) { + return HEXAGON_V32_GET_W(data); + }; + unsigned int UW(void) { + return HEXAGON_V32_GET_UW(data); + }; + + // Extract half word methods + short H0(void) { + return HEXAGON_V32_GET_H0(data); + }; + short H1(void) { + return HEXAGON_V32_GET_H1(data); + }; + unsigned short UH0(void) { + return HEXAGON_V32_GET_UH0(data); + }; + unsigned short UH1(void) { + return HEXAGON_V32_GET_UH1(data); + }; + + // Extract byte methods + signed char B0(void) { + return HEXAGON_V32_GET_B0(data); + }; + signed char B1(void) { + return HEXAGON_V32_GET_B1(data); + }; + signed char B2(void) { + return HEXAGON_V32_GET_B2(data); + }; + signed char B3(void) { + return HEXAGON_V32_GET_B3(data); + }; + unsigned char UB0(void) { + return HEXAGON_V32_GET_UB0(data); + }; + unsigned char UB1(void) { + return HEXAGON_V32_GET_UB1(data); + }; + unsigned char UB2(void) { + return HEXAGON_V32_GET_UB2(data); + }; + unsigned char UB3(void) { + return HEXAGON_V32_GET_UB3(data); + }; + + // NOTE: All set methods return a HEXAGON_Vect32C type + + // Set word method + HEXAGON_Vect32C W(int w) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_W(data, w)); + }; + + // Set half word methods + HEXAGON_Vect32C H0(short h) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_H0(data, h)); + }; + HEXAGON_Vect32C H1(short h) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_H1(data, h)); + }; + + // Set byte methods + HEXAGON_Vect32C B0(signed char b) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_B0(data, b)); + }; + HEXAGON_Vect32C B1(signed char b) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_B1(data, b)); + }; + HEXAGON_Vect32C B2(signed char b) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_B2(data, b)); + }; + HEXAGON_Vect32C B3(signed char b) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_B3(data, b)); + }; + +private: + int data; +}; + +#endif /* __cplusplus */ + +// V65 Silver types +#if __Q6S_ARCH__ >= 65 + // Silver vector types are 128 bytes, and pairs are 256. The vector predicate + // types are 16 bytes and 32 bytes for pairs. + typedef long HEXAGON_VecPred128 __attribute__((__vector_size__(16))) + __attribute__((aligned(128))); + + typedef long HEXAGON_VecPred256 __attribute__((__vector_size__(32))) + __attribute__((aligned(128))); + + typedef long HEXAGON_Vect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + + typedef long HEXAGON_Vect2048 __attribute__((__vector_size__(256))) + __attribute__((aligned(256))); + + typedef long HEXAGON_UVect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(4))); + + typedef long HEXAGON_UVect2048 __attribute__((__vector_size__(256))) + __attribute__((aligned(4))); + + #define Q6S_VectorPredPair HEXAGON_VecPred256 + #define Q6S_VectorPred HEXAGON_VecPred128 + #define Q6S_Vector HEXAGON_Vect1024 + #define Q6S_VectorPair HEXAGON_Vect2048 + #define Q6S_UVector HEXAGON_UVect1024 + #define Q6S_UVectorPair HEXAGON_UVect2048 + +#else /* __Q6S_ARCH__ >= 65 */ + +// V65 Vector types +#if __HVX_ARCH__ >= 65 +#if defined __HVX__ && (__HVX_LENGTH__ == 128) + typedef long HEXAGON_VecPred128 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + + typedef long HEXAGON_Vect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + + typedef long HEXAGON_Vect2048 __attribute__((__vector_size__(256))) + __attribute__((aligned(256))); + + typedef long HEXAGON_UVect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(4))); + + typedef long HEXAGON_UVect2048 __attribute__((__vector_size__(256))) + __attribute__((aligned(4))); + + #define HVX_VectorPred HEXAGON_VecPred128 + #define HVX_Vector HEXAGON_Vect1024 + #define HVX_VectorPair HEXAGON_Vect2048 + #define HVX_UVector HEXAGON_UVect1024 + #define HVX_UVectorPair HEXAGON_UVect2048 +#else /* defined __HVX__ && (__HVX_LENGTH__ == 128) */ +#if defined __HVX__ && (__HVX_LENGTH__ == 64) + typedef long HEXAGON_VecPred64 __attribute__((__vector_size__(64))) + __attribute__((aligned(64))); + + typedef long HEXAGON_Vect512 __attribute__((__vector_size__(64))) + __attribute__((aligned(64))); + + typedef long HEXAGON_Vect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + + typedef long HEXAGON_UVect512 __attribute__((__vector_size__(64))) + __attribute__((aligned(4))); + + typedef long HEXAGON_UVect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(4))); + + #define HVX_VectorPred HEXAGON_VecPred64 + #define HVX_Vector HEXAGON_Vect512 + #define HVX_VectorPair HEXAGON_Vect1024 + #define HVX_UVector HEXAGON_UVect512 + #define HVX_UVectorPair HEXAGON_UVect1024 +#endif /* defined __HVX__ && (__HVX_LENGTH__ == 64) */ +#endif /* defined __HVX__ && (__HVX_LENGTH__ == 128) */ +#endif /* __HVX_ARCH__ >= 65 */ +#endif /* __Q6S_ARCH__ >= 65 */ + +/* Predicates */ + +typedef int HEXAGON_Pred; + +/*** + *** backward compatibility aliases + ***/ + +/* Old names */ +#define Q6Vect Q6Vect64 +#define Q6V_GET_D Q6V64_GET_D +#define Q6V_GET_UD Q6V64_GET_UD +#define Q6V_GET_W0 Q6V64_GET_W0 +#define Q6V_GET_W1 Q6V64_GET_W1 +#define Q6V_GET_UW0 Q6V64_GET_UW0 +#define Q6V_GET_UW1 Q6V64_GET_UW1 +#define Q6V_GET_H0 Q6V64_GET_H0 +#define Q6V_GET_H1 Q6V64_GET_H1 +#define Q6V_GET_H2 Q6V64_GET_H2 +#define Q6V_GET_H3 Q6V64_GET_H3 +#define Q6V_GET_UH0 Q6V64_GET_UH0 +#define Q6V_GET_UH1 Q6V64_GET_UH1 +#define Q6V_GET_UH2 Q6V64_GET_UH2 +#define Q6V_GET_UH3 Q6V64_GET_UH3 +#define Q6V_GET_B0 Q6V64_GET_B0 +#define Q6V_GET_B1 Q6V64_GET_B1 +#define Q6V_GET_B2 Q6V64_GET_B2 +#define Q6V_GET_B3 Q6V64_GET_B3 +#define Q6V_GET_B4 Q6V64_GET_B4 +#define Q6V_GET_B5 Q6V64_GET_B5 +#define Q6V_GET_B6 Q6V64_GET_B6 +#define Q6V_GET_B7 Q6V64_GET_B7 +#define Q6V_GET_UB0 Q6V64_GET_UB0 +#define Q6V_GET_UB1 Q6V64_GET_UB1 +#define Q6V_GET_UB2 Q6V64_GET_UB2 +#define Q6V_GET_UB3 Q6V64_GET_UB3 +#define Q6V_GET_UB4 Q6V64_GET_UB4 +#define Q6V_GET_UB5 Q6V64_GET_UB5 +#define Q6V_GET_UB6 Q6V64_GET_UB6 +#define Q6V_GET_UB7 Q6V64_GET_UB7 +#define Q6V_PUT_D Q6V64_PUT_D +#define Q6V_PUT_W0 Q6V64_PUT_W0 +#define Q6V_PUT_W1 Q6V64_PUT_W1 +#define Q6V_PUT_H0 Q6V64_PUT_H0 +#define Q6V_PUT_H1 Q6V64_PUT_H1 +#define Q6V_PUT_H2 Q6V64_PUT_H2 +#define Q6V_PUT_H3 Q6V64_PUT_H3 +#define Q6V_PUT_B0 Q6V64_PUT_B0 +#define Q6V_PUT_B1 Q6V64_PUT_B1 +#define Q6V_PUT_B2 Q6V64_PUT_B2 +#define Q6V_PUT_B3 Q6V64_PUT_B3 +#define Q6V_PUT_B4 Q6V64_PUT_B4 +#define Q6V_PUT_B5 Q6V64_PUT_B5 +#define Q6V_PUT_B6 Q6V64_PUT_B6 +#define Q6V_PUT_B7 Q6V64_PUT_B7 +#define Q6V_CREATE_D Q6V64_CREATE_D +#define Q6V_CREATE_W Q6V64_CREATE_W +#define Q6V_CREATE_H Q6V64_CREATE_H +#define Q6V_CREATE_B Q6V64_CREATE_B + +#ifdef __cplusplus +#define Q6VectC Q6Vect64C +#endif /* __cplusplus */ + +/* 64 Bit Vectors */ + +typedef long long __attribute__((__may_alias__)) Q6Vect64; + +/* Extract doubleword macros */ + +#define Q6V64_GET_D(v) (v) +#define Q6V64_GET_UD(v) ((unsigned long long)(v)) + +/* Extract word macros */ + +#define Q6V64_GET_W0(v) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.w[0]; \ + }) +#define Q6V64_GET_W1(v) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.w[1]; \ + }) +#define Q6V64_GET_UW0(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned int uw[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.uw[0]; \ + }) +#define Q6V64_GET_UW1(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned int uw[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.uw[1]; \ + }) + +/* Extract half word macros */ + +#define Q6V64_GET_H0(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[0]; \ + }) +#define Q6V64_GET_H1(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[1]; \ + }) +#define Q6V64_GET_H2(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[2]; \ + }) +#define Q6V64_GET_H3(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[3]; \ + }) +#define Q6V64_GET_UH0(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.uh[0]; \ + }) +#define Q6V64_GET_UH1(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.uh[1]; \ + }) +#define Q6V64_GET_UH2(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.uh[2]; \ + }) +#define Q6V64_GET_UH3(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.uh[3]; \ + }) + +/* Extract byte macros */ + +#define Q6V64_GET_B0(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[0]; \ + }) +#define Q6V64_GET_B1(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[1]; \ + }) +#define Q6V64_GET_B2(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[2]; \ + }) +#define Q6V64_GET_B3(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[3]; \ + }) +#define Q6V64_GET_B4(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[4]; \ + }) +#define Q6V64_GET_B5(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[5]; \ + }) +#define Q6V64_GET_B6(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[6]; \ + }) +#define Q6V64_GET_B7(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[7]; \ + }) +#define Q6V64_GET_UB0(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[0]; \ + }) +#define Q6V64_GET_UB1(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[1]; \ + }) +#define Q6V64_GET_UB2(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[2]; \ + }) +#define Q6V64_GET_UB3(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[3]; \ + }) +#define Q6V64_GET_UB4(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[4]; \ + }) +#define Q6V64_GET_UB5(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[5]; \ + }) +#define Q6V64_GET_UB6(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[6]; \ + }) +#define Q6V64_GET_UB7(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[7]; \ + }) + +/* NOTE: All set macros return a Q6Vect64 type */ + +/* Set doubleword macro */ + +#define Q6V64_PUT_D(v, new) (new) + +/* Set word macros */ + +#ifdef __qdsp6__ + +#define Q6V64_PUT_W0(v, new) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.w[0] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_W1(v, new) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.w[1] = (new); \ + _Q6V64_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V64_PUT_W0(v, new) \ + (((v) & 0xffffffff00000000LL) | ((Q6Vect64)((unsigned int)(new)))) +#define Q6V64_PUT_W1(v, new) \ + (((v) & 0x00000000ffffffffLL) | (((Q6Vect64)(new)) << 32LL)) + +#endif /* !__qdsp6__ */ + +/* Set half word macros */ + +#ifdef __qdsp6__ + +#define Q6V64_PUT_H0(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[0] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_H1(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[1] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_H2(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[2] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_H3(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[3] = (new); \ + _Q6V64_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V64_PUT_H0(v, new) \ + (((v) & 0xffffffffffff0000LL) | ((Q6Vect64)((unsigned short)(new)))) +#define Q6V64_PUT_H1(v, new) \ + (((v) & 0xffffffff0000ffffLL) | (((Q6Vect64)((unsigned short)(new))) << 16LL)) +#define Q6V64_PUT_H2(v, new) \ + (((v) & 0xffff0000ffffffffLL) | (((Q6Vect64)((unsigned short)(new))) << 32LL)) +#define Q6V64_PUT_H3(v, new) \ + (((v) & 0x0000ffffffffffffLL) | (((Q6Vect64)(new)) << 48LL)) + +#endif /* !__qdsp6__ */ + +/* Set byte macros */ + +#ifdef __qdsp6__ + +#define Q6V64_PUT_B0(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[0] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B1(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[1] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B2(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[2] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B3(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[3] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B4(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[4] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B5(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[5] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B6(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[6] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B7(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[7] = (new); \ + _Q6V64_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V64_PUT_B0(v, new) \ + (((v) & 0xffffffffffffff00LL) | ((Q6Vect64)((unsigned char)(new)))) +#define Q6V64_PUT_B1(v, new) \ + (((v) & 0xffffffffffff00ffLL) | (((Q6Vect64)((unsigned char)(new))) << 8LL)) +#define Q6V64_PUT_B2(v, new) \ + (((v) & 0xffffffffff00ffffLL) | (((Q6Vect64)((unsigned char)(new))) << 16LL)) +#define Q6V64_PUT_B3(v, new) \ + (((v) & 0xffffffff00ffffffLL) | (((Q6Vect64)((unsigned char)(new))) << 24LL)) +#define Q6V64_PUT_B4(v, new) \ + (((v) & 0xffffff00ffffffffLL) | (((Q6Vect64)((unsigned char)(new))) << 32LL)) +#define Q6V64_PUT_B5(v, new) \ + (((v) & 0xffff00ffffffffffLL) | (((Q6Vect64)((unsigned char)(new))) << 40LL)) +#define Q6V64_PUT_B6(v, new) \ + (((v) & 0xff00ffffffffffffLL) | (((Q6Vect64)((unsigned char)(new))) << 48LL)) +#define Q6V64_PUT_B7(v, new) \ + (((v) & 0x00ffffffffffffffLL) | (((Q6Vect64)(new)) << 56LL)) + +#endif /* !__qdsp6__ */ + +/* NOTE: All create macros return a Q6Vect64 type */ + +/* Create from a doubleword */ + +#define Q6V64_CREATE_D(d) (d) + +/* Create from words */ + +#ifdef __qdsp6__ + +#define Q6V64_CREATE_W(w1, w0) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.w[0] = (w0); \ + _Q6V64_internal_union.w[1] = (w1); \ + _Q6V64_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V64_CREATE_W(w1, w0) \ + ((((Q6Vect64)(w1)) << 32LL) | ((Q6Vect64)((w0) & 0xffffffff))) + +#endif /* !__qdsp6__ */ + +/* Create from half words */ + +#ifdef __qdsp6__ + +#define Q6V64_CREATE_H(h3, h2, h1, h0) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.h[0] = (h0); \ + _Q6V64_internal_union.h[1] = (h1); \ + _Q6V64_internal_union.h[2] = (h2); \ + _Q6V64_internal_union.h[3] = (h3); \ + _Q6V64_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V64_CREATE_H(h3, h2, h1, h0) \ + ((((Q6Vect64)(h3)) << 48LL) | (((Q6Vect64)((h2) & 0xffff)) << 32LL) | \ + (((Q6Vect64)((h1) & 0xffff)) << 16LL) | ((Q6Vect64)((h0) & 0xffff))) + +#endif /* !__qdsp6__ */ + +/* Create from bytes */ + +#ifdef __qdsp6__ + +#define Q6V64_CREATE_B(b7, b6, b5, b4, b3, b2, b1, b0) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.b[0] = (b0); \ + _Q6V64_internal_union.b[1] = (b1); \ + _Q6V64_internal_union.b[2] = (b2); \ + _Q6V64_internal_union.b[3] = (b3); \ + _Q6V64_internal_union.b[4] = (b4); \ + _Q6V64_internal_union.b[5] = (b5); \ + _Q6V64_internal_union.b[6] = (b6); \ + _Q6V64_internal_union.b[7] = (b7); \ + _Q6V64_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V64_CREATE_B(b7, b6, b5, b4, b3, b2, b1, b0) \ + ((((Q6Vect64)(b7)) << 56LL) | (((Q6Vect64)((b6) & 0xff)) << 48LL) | \ + (((Q6Vect64)((b5) & 0xff)) << 40LL) | (((Q6Vect64)((b4) & 0xff)) << 32LL) | \ + (((Q6Vect64)((b3) & 0xff)) << 24LL) | (((Q6Vect64)((b2) & 0xff)) << 16LL) | \ + (((Q6Vect64)((b1) & 0xff)) << 8LL) | ((Q6Vect64)((b0) & 0xff))) + +#endif /* !__qdsp6__ */ + +#ifdef __cplusplus + +class Q6Vect64C { +public: + // Constructors + Q6Vect64C(long long d = 0) : data(d) {}; + Q6Vect64C(int w1, int w0) : data(Q6V64_CREATE_W(w1, w0)) {}; + Q6Vect64C(short h3, short h2, short h1, short h0) + : data(Q6V64_CREATE_H(h3, h2, h1, h0)) {}; + Q6Vect64C(signed char b7, signed char b6, signed char b5, signed char b4, + signed char b3, signed char b2, signed char b1, signed char b0) + : data(Q6V64_CREATE_B(b7, b6, b5, b4, b3, b2, b1, b0)) {}; + Q6Vect64C(const Q6Vect64C &v) : data(v.data) {}; + + Q6Vect64C &operator=(const Q6Vect64C &v) { + data = v.data; + return *this; + }; + + operator long long() { + return data; + }; + + // Extract doubleword methods + long long D(void) { + return Q6V64_GET_D(data); + }; + unsigned long long UD(void) { + return Q6V64_GET_UD(data); + }; + + // Extract word methods + int W0(void) { + return Q6V64_GET_W0(data); + }; + int W1(void) { + return Q6V64_GET_W1(data); + }; + unsigned int UW0(void) { + return Q6V64_GET_UW0(data); + }; + unsigned int UW1(void) { + return Q6V64_GET_UW1(data); + }; + + // Extract half word methods + short H0(void) { + return Q6V64_GET_H0(data); + }; + short H1(void) { + return Q6V64_GET_H1(data); + }; + short H2(void) { + return Q6V64_GET_H2(data); + }; + short H3(void) { + return Q6V64_GET_H3(data); + }; + unsigned short UH0(void) { + return Q6V64_GET_UH0(data); + }; + unsigned short UH1(void) { + return Q6V64_GET_UH1(data); + }; + unsigned short UH2(void) { + return Q6V64_GET_UH2(data); + }; + unsigned short UH3(void) { + return Q6V64_GET_UH3(data); + }; + + // Extract byte methods + signed char B0(void) { + return Q6V64_GET_B0(data); + }; + signed char B1(void) { + return Q6V64_GET_B1(data); + }; + signed char B2(void) { + return Q6V64_GET_B2(data); + }; + signed char B3(void) { + return Q6V64_GET_B3(data); + }; + signed char B4(void) { + return Q6V64_GET_B4(data); + }; + signed char B5(void) { + return Q6V64_GET_B5(data); + }; + signed char B6(void) { + return Q6V64_GET_B6(data); + }; + signed char B7(void) { + return Q6V64_GET_B7(data); + }; + unsigned char UB0(void) { + return Q6V64_GET_UB0(data); + }; + unsigned char UB1(void) { + return Q6V64_GET_UB1(data); + }; + unsigned char UB2(void) { + return Q6V64_GET_UB2(data); + }; + unsigned char UB3(void) { + return Q6V64_GET_UB3(data); + }; + unsigned char UB4(void) { + return Q6V64_GET_UB4(data); + }; + unsigned char UB5(void) { + return Q6V64_GET_UB5(data); + }; + unsigned char UB6(void) { + return Q6V64_GET_UB6(data); + }; + unsigned char UB7(void) { + return Q6V64_GET_UB7(data); + }; + + // NOTE: All set methods return a Q6Vect64C type + + // Set doubleword method + Q6Vect64C D(long long d) { + return Q6Vect64C(Q6V64_PUT_D(data, d)); + }; + + // Set word methods + Q6Vect64C W0(int w) { + return Q6Vect64C(Q6V64_PUT_W0(data, w)); + }; + Q6Vect64C W1(int w) { + return Q6Vect64C(Q6V64_PUT_W1(data, w)); + }; + + // Set half word methods + Q6Vect64C H0(short h) { + return Q6Vect64C(Q6V64_PUT_H0(data, h)); + }; + Q6Vect64C H1(short h) { + return Q6Vect64C(Q6V64_PUT_H1(data, h)); + }; + Q6Vect64C H2(short h) { + return Q6Vect64C(Q6V64_PUT_H2(data, h)); + }; + Q6Vect64C H3(short h) { + return Q6Vect64C(Q6V64_PUT_H3(data, h)); + }; + + // Set byte methods + Q6Vect64C B0(signed char b) { + return Q6Vect64C(Q6V64_PUT_B0(data, b)); + }; + Q6Vect64C B1(signed char b) { + return Q6Vect64C(Q6V64_PUT_B1(data, b)); + }; + Q6Vect64C B2(signed char b) { + return Q6Vect64C(Q6V64_PUT_B2(data, b)); + }; + Q6Vect64C B3(signed char b) { + return Q6Vect64C(Q6V64_PUT_B3(data, b)); + }; + Q6Vect64C B4(signed char b) { + return Q6Vect64C(Q6V64_PUT_B4(data, b)); + }; + Q6Vect64C B5(signed char b) { + return Q6Vect64C(Q6V64_PUT_B5(data, b)); + }; + Q6Vect64C B6(signed char b) { + return Q6Vect64C(Q6V64_PUT_B6(data, b)); + }; + Q6Vect64C B7(signed char b) { + return Q6Vect64C(Q6V64_PUT_B7(data, b)); + }; + +private: + long long data; +}; + +#endif /* __cplusplus */ + +/* 32 Bit Vectors */ + +typedef int Q6Vect32; + +/* Extract word macros */ + +#define Q6V32_GET_W(v) (v) +#define Q6V32_GET_UW(v) ((unsigned int)(v)) + +/* Extract half word macros */ + +#define Q6V32_GET_H0(v) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.h[0]; \ + }) +#define Q6V32_GET_H1(v) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.h[1]; \ + }) +#define Q6V32_GET_UH0(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned short uh[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.uh[0]; \ + }) +#define Q6V32_GET_UH1(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned short uh[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.uh[1]; \ + }) + +/* Extract byte macros */ + +#define Q6V32_GET_B0(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[0]; \ + }) +#define Q6V32_GET_B1(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[1]; \ + }) +#define Q6V32_GET_B2(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[2]; \ + }) +#define Q6V32_GET_B3(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[3]; \ + }) +#define Q6V32_GET_UB0(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.ub[0]; \ + }) +#define Q6V32_GET_UB1(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.ub[1]; \ + }) +#define Q6V32_GET_UB2(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.ub[2]; \ + }) +#define Q6V32_GET_UB3(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.ub[3]; \ + }) + +/* NOTE: All set macros return a Q6Vect32 type */ + +/* Set word macro */ + +#define Q6V32_PUT_W(v, new) (new) + +/* Set half word macros */ + +#ifdef __qdsp6__ + +#define Q6V32_PUT_H0(v, new) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.h[0] = (new); \ + _Q6V32_internal_union.w; \ + }) +#define Q6V32_PUT_H1(v, new) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.h[1] = (new); \ + _Q6V32_internal_union.w; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V32_PUT_H0(v, new) \ + (((v) & 0xffff0000) | ((Q6Vect32)((unsigned short)(new)))) +#define Q6V32_PUT_H1(v, new) (((v) & 0x0000ffff) | (((Q6Vect32)(new)) << 16)) + +#endif /* !__qdsp6__ */ + +/* Set byte macros */ + +#ifdef __qdsp6__ + +#define Q6V32_PUT_B0(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[0] = (new); \ + _Q6V32_internal_union.w; \ + }) +#define Q6V32_PUT_B1(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[1] = (new); \ + _Q6V32_internal_union.w; \ + }) +#define Q6V32_PUT_B2(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[2] = (new); \ + _Q6V32_internal_union.w; \ + }) +#define Q6V32_PUT_B3(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[3] = (new); \ + _Q6V32_internal_union.w; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V32_PUT_B0(v, new) \ + (((v) & 0xffffff00) | ((Q6Vect32)((unsigned char)(new)))) +#define Q6V32_PUT_B1(v, new) \ + (((v) & 0xffff00ff) | (((Q6Vect32)((unsigned char)(new))) << 8)) +#define Q6V32_PUT_B2(v, new) \ + (((v) & 0xff00ffff) | (((Q6Vect32)((unsigned char)(new))) << 16)) +#define Q6V32_PUT_B3(v, new) (((v) & 0x00ffffff) | (((Q6Vect32)(new)) << 24)) + +#endif /* !__qdsp6__ */ + +/* NOTE: All create macros return a Q6Vect32 type */ + +/* Create from a word */ + +#define Q6V32_CREATE_W(w) (w) + +/* Create from half words */ + +#ifdef __qdsp6__ + +#define Q6V32_CREATE_H(h1, h0) \ + __extension__({ \ + union { \ + long long d; \ + short h[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.h[0] = (h0); \ + _Q6V32_internal_union.h[1] = (h1); \ + _Q6V32_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V32_CREATE_H(h1, h0) \ + ((((Q6Vect32)(h1)) << 16) | ((Q6Vect32)((h0) & 0xffff))) + +#endif /* !__qdsp6__ */ + +/* Create from bytes */ +#ifdef __qdsp6__ + +#define Q6V32_CREATE_B(b3, b2, b1, b0) \ + __extension__({ \ + union { \ + long long d; \ + char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.b[0] = (b0); \ + _Q6V32_internal_union.b[1] = (b1); \ + _Q6V32_internal_union.b[2] = (b2); \ + _Q6V32_internal_union.b[3] = (b3); \ + _Q6V32_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V32_CREATE_B(b3, b2, b1, b0) \ + ((((Q6Vect32)(b3)) << 24) | (((Q6Vect32)((b2) & 0xff)) << 16) | \ + (((Q6Vect32)((b1) & 0xff)) << 8) | ((Q6Vect32)((b0) & 0xff))) + +#endif /* !__qdsp6__ */ + +#ifdef __cplusplus + +class Q6Vect32C { +public: + // Constructors + Q6Vect32C(int w = 0) : data(w) {}; + Q6Vect32C(short h1, short h0) : data(Q6V32_CREATE_H(h1, h0)) {}; + Q6Vect32C(signed char b3, signed char b2, signed char b1, signed char b0) + : data(Q6V32_CREATE_B(b3, b2, b1, b0)) {}; + Q6Vect32C(const Q6Vect32C &v) : data(v.data) {}; + + Q6Vect32C &operator=(const Q6Vect32C &v) { + data = v.data; + return *this; + }; + + operator int() { + return data; + }; + + // Extract word methods + int W(void) { + return Q6V32_GET_W(data); + }; + unsigned int UW(void) { + return Q6V32_GET_UW(data); + }; + + // Extract half word methods + short H0(void) { + return Q6V32_GET_H0(data); + }; + short H1(void) { + return Q6V32_GET_H1(data); + }; + unsigned short UH0(void) { + return Q6V32_GET_UH0(data); + }; + unsigned short UH1(void) { + return Q6V32_GET_UH1(data); + }; + + // Extract byte methods + signed char B0(void) { + return Q6V32_GET_B0(data); + }; + signed char B1(void) { + return Q6V32_GET_B1(data); + }; + signed char B2(void) { + return Q6V32_GET_B2(data); + }; + signed char B3(void) { + return Q6V32_GET_B3(data); + }; + unsigned char UB0(void) { + return Q6V32_GET_UB0(data); + }; + unsigned char UB1(void) { + return Q6V32_GET_UB1(data); + }; + unsigned char UB2(void) { + return Q6V32_GET_UB2(data); + }; + unsigned char UB3(void) { + return Q6V32_GET_UB3(data); + }; + + // NOTE: All set methods return a Q6Vect32C type + + // Set word method + Q6Vect32C W(int w) { + return Q6Vect32C(Q6V32_PUT_W(data, w)); + }; + + // Set half word methods + Q6Vect32C H0(short h) { + return Q6Vect32C(Q6V32_PUT_H0(data, h)); + }; + Q6Vect32C H1(short h) { + return Q6Vect32C(Q6V32_PUT_H1(data, h)); + }; + + // Set byte methods + Q6Vect32C B0(signed char b) { + return Q6Vect32C(Q6V32_PUT_B0(data, b)); + }; + Q6Vect32C B1(signed char b) { + return Q6Vect32C(Q6V32_PUT_B1(data, b)); + }; + Q6Vect32C B2(signed char b) { + return Q6Vect32C(Q6V32_PUT_B2(data, b)); + }; + Q6Vect32C B3(signed char b) { + return Q6Vect32C(Q6V32_PUT_B3(data, b)); + }; + +private: + int data; +}; + +#endif /* __cplusplus */ + +// V65 Vector types +#if __HVX_ARCH__ >= 65 +#if defined __HVX__ && (__HVX_LENGTH__ == 128) +typedef long Q6VecPred128 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + +typedef long Q6Vect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + +typedef long Q6Vect2048 __attribute__((__vector_size__(256))) + __attribute__((aligned(256))); + +#else /* defined __HVX__ && (__HVX_LENGTH__ == 128) */ +#if defined __HVX__ && (__HVX_LENGTH__ == 64) +typedef long Q6VecPred64 __attribute__((__vector_size__(64))) + __attribute__((aligned(64))); + +typedef long Q6Vect512 __attribute__((__vector_size__(64))) + __attribute__((aligned(64))); + +typedef long Q6Vect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + +#endif /* defined __HVX__ && (__HVX_LENGTH__ == 64) */ +#endif /* defined __HVX__ && (__HVX_LENGTH__ == 128) */ +#endif /* __HVX_ARCH__ >= 65 */ + +/* Predicates */ + +typedef int Q6Pred; + + +#ifdef __HVX__ + +// Extract HVX VectorPair macro. +#define HEXAGON_HVX_GET_W(v) (v) + +// Extract HVX Vector macros. +#define HEXAGON_HVX_GET_V0(v) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_Vector V[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.W = (v); \ + _HEXAGON_HVX_internal_union.V[0]; \ + }) +#define HEXAGON_HVX_GET_V1(v) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_Vector V[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.W = (v); \ + _HEXAGON_HVX_internal_union.V[1]; \ + }) +#define HEXAGON_HVX_GET_P(v) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_VectorPred P[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.W = (v); \ + _HEXAGON_HVX_internal_union.P[0]; \ + }) + +// Set HVX VectorPair macro. +#define HEXAGON_HVX_PUT_W(v, new) (new) + +// Set HVX Vector macros. +#define HEXAGON_HVX_PUT_V0(v, new) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_Vector V[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.W = (v); \ + _HEXAGON_HVX_internal_union.V[0] = (new); \ + _HEXAGON_HVX_internal_union.W; \ + }) + +#define HEXAGON_HVX_PUT_V1(v, new) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_Vector V[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.W = (v); \ + _HEXAGON_HVX_internal_union.V[1] = (new); \ + _HEXAGON_HVX_internal_union.W; \ + }) + +#define HEXAGON_HVX_PUT_P(v, new) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_VectorPred P[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.W = (v); \ + _HEXAGON_HVX_internal_union.P[0] = (new); \ + _HEXAGON_HVX_internal_union.W; \ + }) + + +#define HEXAGON_HVX_CREATE_W(v1, v0) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_Vector V[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.V[0] = (v0); \ + _HEXAGON_HVX_internal_union.V[1] = (v1); \ + _HEXAGON_HVX_internal_union.W; \ + }) + +#ifdef __cplusplus + +class HVX_Vect { +public: + // Constructors. + // Default. + HVX_Vect() : data(Q6_W_vcombine_VV(Q6_V_vzero(), Q6_V_vzero())){}; + + // Custom constructors. + HVX_Vect(HVX_VectorPair W) : data(W){}; + HVX_Vect(HVX_Vector v1, HVX_Vector v0) : data(HEXAGON_HVX_CREATE_W(v1, v0)){}; + + // Copy constructor. + HVX_Vect(const HVX_Vect &W) = default; + + // Move constructor. + HVX_Vect(HVX_Vect &&W) = default; + + // Assignment operator. + HVX_Vect &operator=(const HVX_Vect &W) = default; + + operator HVX_VectorPair() { return data; }; + + // Extract VectorPair method. + HVX_VectorPair W(void) { return HEXAGON_HVX_GET_W(data); }; + + // Extract Vector methods. + HVX_Vector V0(void) { return HEXAGON_HVX_GET_V0(data); }; + HVX_Vector V1(void) { return HEXAGON_HVX_GET_V1(data); }; + HVX_VectorPred P(void) { return HEXAGON_HVX_GET_P(data); }; + + // NOTE: All set methods return a HVX_Vect type. + // Set HVX VectorPair method. + HVX_Vect W(HVX_VectorPair w) { return HVX_Vect(HEXAGON_HVX_PUT_W(data, w)); }; + + // Set HVX Vector methods. + HVX_Vect V0(HVX_Vector v) { return HVX_Vect(HEXAGON_HVX_PUT_V0(data, v)); }; + HVX_Vect V1(HVX_Vector v) { return HVX_Vect(HEXAGON_HVX_PUT_V1(data, v)); }; + HVX_Vect P(HVX_VectorPred p) { return HVX_Vect(HEXAGON_HVX_PUT_P(data, p)); }; + +private: + HVX_VectorPair data; +}; + +#endif /* __cplusplus */ +#endif /* __HVX__ */ + +#define HEXAGON_UDMA_DM0_STATUS_IDLE 0x00000000 +#define HEXAGON_UDMA_DM0_STATUS_RUN 0x00000001 +#define HEXAGON_UDMA_DM0_STATUS_ERROR 0x00000002 +#define HEXAGON_UDMA_DESC_DSTATE_INCOMPLETE 0 +#define HEXAGON_UDMA_DESC_DSTATE_COMPLETE 1 +#define HEXAGON_UDMA_DESC_ORDER_NOORDER 0 +#define HEXAGON_UDMA_DESC_ORDER_ORDER 1 +#define HEXAGON_UDMA_DESC_BYPASS_OFF 0 +#define HEXAGON_UDMA_DESC_BYPASS_ON 1 +#define HEXAGON_UDMA_DESC_COMP_NONE 0 +#define HEXAGON_UDMA_DESC_COMP_DLBC 1 +#define HEXAGON_UDMA_DESC_DESCTYPE_TYPE0 0 +#define HEXAGON_UDMA_DESC_DESCTYPE_TYPE1 1 + +typedef struct hexagon_udma_descriptor_type0_s +{ + void *next; + unsigned int length:24; + unsigned int desctype:2; + unsigned int dstcomp:1; + unsigned int srccomp:1; + unsigned int dstbypass:1; + unsigned int srcbypass:1; + unsigned int order:1; + unsigned int dstate:1; + void *src; + void *dst; +} hexagon_udma_descriptor_type0_t; + +typedef struct hexagon_udma_descriptor_type1_s +{ + void *next; + unsigned int length:24; + unsigned int desctype:2; + unsigned int dstcomp:1; + unsigned int srccomp:1; + unsigned int dstbypass:1; + unsigned int srcbypass:1; + unsigned int order:1; + unsigned int dstate:1; + void *src; + void *dst; + unsigned int allocation:28; + unsigned int padding:4; + unsigned int roiwidth:16; + unsigned int roiheight:16; + unsigned int srcstride:16; + unsigned int dststride:16; + unsigned int srcwidthoffset:16; + unsigned int dstwidthoffset:16; +} hexagon_udma_descriptor_type1_t; + +#endif /* !HEXAGON_TYPES_H */ diff --git a/lib/include/hvx_hexagon_protos.h b/lib/include/hvx_hexagon_protos.h new file mode 100644 index 0000000000..41ce7a6b93 --- /dev/null +++ b/lib/include/hvx_hexagon_protos.h @@ -0,0 +1,4392 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +// Automatically generated file, do not edit! +//===----------------------------------------------------------------------===// + + + +#ifndef _HVX_HEXAGON_PROTOS_H_ +#define _HVX_HEXAGON_PROTOS_H_ 1 + +#ifdef __HVX__ +#if __HVX_LENGTH__ == 128 +#define __BUILTIN_VECTOR_WRAP(a) a ## _128B +#else +#define __BUILTIN_VECTOR_WRAP(a) a +#endif + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rd32=vextract(Vu32,Rs32) + C Intrinsic Prototype: Word32 Q6_R_vextract_VR(HVX_Vector Vu, Word32 Rs) + Instruction Type: LD + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_R_vextract_VR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_extractw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=hi(Vss32) + C Intrinsic Prototype: HVX_Vector Q6_V_hi_W(HVX_VectorPair Vss) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_hi_W __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_hi) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=lo(Vss32) + C Intrinsic Prototype: HVX_Vector Q6_V_lo_W(HVX_VectorPair Vss) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_lo_W __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_lo) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vsplat(Rt32) + C Intrinsic Prototype: HVX_Vector Q6_V_vsplat_R(Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_V_vsplat_R __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_lvsplatw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=and(Qs4,Qt4) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_and_QQ(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_and_QQ __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=and(Qs4,!Qt4) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_and_QQn(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_and_QQn __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_and_n) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=not(Qs4) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_not_Q(HVX_VectorPred Qs) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_not_Q __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_not) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=or(Qs4,Qt4) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_or_QQ(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_or_QQ __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=or(Qs4,!Qt4) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_or_QQn(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_or_QQn __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_or_n) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vsetq(Rt32) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vsetq_R(Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vsetq_R __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_scalar2) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=xor(Qs4,Qt4) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_xor_QQ(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_xor_QQ __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) vmem(Rt32+#s4)=Vs32 + C Intrinsic Prototype: void Q6_vmem_QnRIV(HVX_VectorPred Qv, HVX_Vector* Rt, HVX_Vector Vs) + Instruction Type: CVI_VM_ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmem_QnRIV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vS32b_nqpred_ai) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) vmem(Rt32+#s4):nt=Vs32 + C Intrinsic Prototype: void Q6_vmem_QnRIV_nt(HVX_VectorPred Qv, HVX_Vector* Rt, HVX_Vector Vs) + Instruction Type: CVI_VM_ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmem_QnRIV_nt __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vS32b_nt_nqpred_ai) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) vmem(Rt32+#s4):nt=Vs32 + C Intrinsic Prototype: void Q6_vmem_QRIV_nt(HVX_VectorPred Qv, HVX_Vector* Rt, HVX_Vector Vs) + Instruction Type: CVI_VM_ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmem_QRIV_nt __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vS32b_nt_qpred_ai) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) vmem(Rt32+#s4)=Vs32 + C Intrinsic Prototype: void Q6_vmem_QRIV(HVX_VectorPred Qv, HVX_Vector* Rt, HVX_Vector Vs) + Instruction Type: CVI_VM_ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmem_QRIV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vS32b_qpred_ai) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vabsdiff(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vabsdiff_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuh_vabsdiff_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsdiffh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vabsdiff(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vub_vabsdiff_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vub_vabsdiff_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsdiffub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vabsdiff(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vabsdiff_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuh_vabsdiff_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsdiffuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vabsdiff(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vabsdiff_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vabsdiff_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsdiffw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vabs(Vu32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vabs_Vh(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vabs_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vabs(Vu32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vabs_Vh_sat(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vabs_Vh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsh_sat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vabs(Vu32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vabs_Vw(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vabs_Vw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vabs(Vu32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vabs_Vw_sat(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vabs_Vw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsw_sat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vadd(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vadd_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vadd_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.b=vadd(Vuu32.b,Vvv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wb_vadd_WbWb(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wb_vadd_WbWb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddb_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) Vx32.b+=Vu32.b + C Intrinsic Prototype: HVX_Vector Q6_Vb_condacc_QnVbVb(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_condacc_QnVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddbnq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) Vx32.b+=Vu32.b + C Intrinsic Prototype: HVX_Vector Q6_Vb_condacc_QVbVb(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_condacc_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddbq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vadd(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vadd_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vadd_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vadd(Vuu32.h,Vvv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vadd_WhWh(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vadd_WhWh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddh_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) Vx32.h+=Vu32.h + C Intrinsic Prototype: HVX_Vector Q6_Vh_condacc_QnVhVh(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_condacc_QnVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddhnq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) Vx32.h+=Vu32.h + C Intrinsic Prototype: HVX_Vector Q6_Vh_condacc_QVhVh(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_condacc_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddhq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vadd(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vadd_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vadd_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vadd(Vuu32.h,Vvv32.h):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vadd_WhWh_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vadd_WhWh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddhsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vadd(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vadd_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vadd_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddhw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vadd(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vadd_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vadd_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddubh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vadd(Vu32.ub,Vv32.ub):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vadd_VubVub_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vadd_VubVub_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddubsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.ub=vadd(Vuu32.ub,Vvv32.ub):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wub_vadd_WubWub_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wub_vadd_WubWub_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddubsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vadd(Vu32.uh,Vv32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vadd_VuhVuh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vadd_VuhVuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vadduhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uh=vadd(Vuu32.uh,Vvv32.uh):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vadd_WuhWuh_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuh_vadd_WuhWuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vadduhsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vadd(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vadd_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vadd_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vadduhw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vadd(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vadd_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vadd_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vadd(Vuu32.w,Vvv32.w) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vadd_WwWw(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vadd_WwWw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddw_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) Vx32.w+=Vu32.w + C Intrinsic Prototype: HVX_Vector Q6_Vw_condacc_QnVwVw(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_condacc_QnVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddwnq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) Vx32.w+=Vu32.w + C Intrinsic Prototype: HVX_Vector Q6_Vw_condacc_QVwVw(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_condacc_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddwq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vadd(Vu32.w,Vv32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vadd_VwVw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vadd_VwVw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddwsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vadd(Vuu32.w,Vvv32.w):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vadd_WwWw_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vadd_WwWw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddwsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=valign(Vu32,Vv32,Rt8) + C Intrinsic Prototype: HVX_Vector Q6_V_valign_VVR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_valign_VVR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_valignb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=valign(Vu32,Vv32,#u3) + C Intrinsic Prototype: HVX_Vector Q6_V_valign_VVI(HVX_Vector Vu, HVX_Vector Vv, Word32 Iu3) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_valign_VVI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_valignbi) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vand(Vu32,Vv32) + C Intrinsic Prototype: HVX_Vector Q6_V_vand_VV(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vand_VV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vand) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vand(Qu4,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_V_vand_QR(HVX_VectorPred Qu, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_V_vand_QR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandqrt) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32|=vand(Qu4,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_V_vandor_VQR(HVX_Vector Vx, HVX_VectorPred Qu, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_V_vandor_VQR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandqrt_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vand(Vu32,Rt32) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vand_VR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Q_vand_VR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandvrt) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vand(Vu32,Rt32) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vandor_QVR(HVX_VectorPred Qx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Q_vandor_QVR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandvrt_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasl(Vu32.h,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasl_VhR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasl_VhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaslh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasl(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasl_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasl_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaslhv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vasl(Vu32.w,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vasl_VwR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vasl_VwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaslw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vasl(Vu32.w,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vaslacc_VwVwR(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vaslacc_VwVwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaslw_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vasl(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vasl_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vasl_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaslwv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasr(Vu32.h,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasr_VhR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasr_VhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vasr(Vu32.h,Vv32.h,Rt8):rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vasr_VhVhR_rnd_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vasr_VhVhR_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrhbrndsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vasr(Vu32.h,Vv32.h,Rt8):rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vasr_VhVhR_rnd_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vasr_VhVhR_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrhubrndsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vasr(Vu32.h,Vv32.h,Rt8):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vasr_VhVhR_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vasr_VhVhR_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrhubsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasr(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasr_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasr_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrhv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vasr(Vu32.w,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vasr_VwR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vasr_VwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vasr(Vu32.w,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vasracc_VwVwR(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vasracc_VwVwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrw_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasr(Vu32.w,Vv32.w,Rt8) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasr_VwVwR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasr_VwVwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrwh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasr(Vu32.w,Vv32.w,Rt8):rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasr_VwVwR_rnd_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasr_VwVwR_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrwhrndsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasr(Vu32.w,Vv32.w,Rt8):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasr_VwVwR_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasr_VwVwR_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrwhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vasr(Vu32.w,Vv32.w,Rt8):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vasr_VwVwR_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vasr_VwVwR_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrwuhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vasr(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vasr_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vasr_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrwv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=Vu32 + C Intrinsic Prototype: HVX_Vector Q6_V_equals_V(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_equals_V __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vassign) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32=Vuu32 + C Intrinsic Prototype: HVX_VectorPair Q6_W_equals_W(HVX_VectorPair Vuu) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_W_equals_W __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vassignp) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vavg(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vavg_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vavg_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vavg(Vu32.h,Vv32.h):rnd + C Intrinsic Prototype: HVX_Vector Q6_Vh_vavg_VhVh_rnd(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vavg_VhVh_rnd __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavghrnd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vavg(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vub_vavg_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vavg_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vavg(Vu32.ub,Vv32.ub):rnd + C Intrinsic Prototype: HVX_Vector Q6_Vub_vavg_VubVub_rnd(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vavg_VubVub_rnd __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgubrnd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vavg(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vavg_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vavg_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavguh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vavg(Vu32.uh,Vv32.uh):rnd + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vavg_VuhVuh_rnd(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vavg_VuhVuh_rnd __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavguhrnd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vavg(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vavg_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vavg_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vavg(Vu32.w,Vv32.w):rnd + C Intrinsic Prototype: HVX_Vector Q6_Vw_vavg_VwVw_rnd(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vavg_VwVw_rnd __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgwrnd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vcl0(Vu32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vcl0_Vuh(HVX_Vector Vu) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vcl0_Vuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vcl0h) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vcl0(Vu32.uw) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vcl0_Vuw(HVX_Vector Vu) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vcl0_Vuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vcl0w) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32=vcombine(Vu32,Vv32) + C Intrinsic Prototype: HVX_VectorPair Q6_W_vcombine_VV(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_W_vcombine_VV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vcombine) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=#0 + C Intrinsic Prototype: HVX_Vector Q6_V_vzero() + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vzero __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vd0) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vdeal(Vu32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vdeal_Vb(HVX_Vector Vu) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vdeal_Vb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdealb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vdeale(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vdeale_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vdeale_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdealb4w) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vdeal(Vu32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vdeal_Vh(HVX_Vector Vu) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vdeal_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdealh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32=vdeal(Vu32,Vv32,Rt8) + C Intrinsic Prototype: HVX_VectorPair Q6_W_vdeal_VVR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_W_vdeal_VVR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdealvdd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vdelta(Vu32,Vv32) + C Intrinsic Prototype: HVX_Vector Q6_V_vdelta_VV(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vdelta_VV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdelta) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vdmpy(Vu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vdmpy_VubRb(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vdmpy_VubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpybus) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.h+=vdmpy(Vu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vdmpyacc_VhVubRb(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vdmpyacc_VhVubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpybus_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vdmpy(Vuu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vdmpy_WubRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vdmpy_WubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpybus_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vdmpy(Vuu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vdmpyacc_WhWubRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vdmpyacc_WhWubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpybus_dv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vdmpy(Vu32.h,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpy_VhRb(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpy_VhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vdmpy(Vu32.h,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpyacc_VwVhRb(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpyacc_VwVhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhb_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vdmpy(Vuu32.h,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vdmpy_WhRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vdmpy_WhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhb_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vdmpy(Vuu32.h,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vdmpyacc_WwWhRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vdmpyacc_WwWhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhb_dv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vdmpy(Vuu32.h,Rt32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpy_WhRh_sat(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpy_WhRh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhisat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vdmpy(Vuu32.h,Rt32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpyacc_VwWhRh_sat(HVX_Vector Vx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpyacc_VwWhRh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhisat_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vdmpy(Vu32.h,Rt32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpy_VhRh_sat(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpy_VhRh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vdmpy(Vu32.h,Rt32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpyacc_VwVhRh_sat(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpyacc_VwVhRh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhsat_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vdmpy(Vuu32.h,Rt32.uh,#1):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpy_WhRuh_sat(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpy_WhRuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhsuisat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vdmpy(Vuu32.h,Rt32.uh,#1):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpyacc_VwWhRuh_sat(HVX_Vector Vx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpyacc_VwWhRuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhsuisat_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vdmpy(Vu32.h,Rt32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpy_VhRuh_sat(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpy_VhRuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhsusat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vdmpy(Vu32.h,Rt32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpyacc_VwVhRuh_sat(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpyacc_VwVhRuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhsusat_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vdmpy(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpy_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpy_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhvsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vdmpy(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpyacc_VwVhVh_sat(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpyacc_VwVhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhvsat_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vdsad(Vuu32.uh,Rt32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vdsad_WuhRuh(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vdsad_WuhRuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdsaduh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uw+=vdsad(Vuu32.uh,Rt32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vdsadacc_WuwWuhRuh(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vdsadacc_WuwWuhRuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdsaduh_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.eq(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eq_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eq_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.eq(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqand_QVbVb(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqand_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqb_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.eq(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqor_QVbVb(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqor_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqb_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.eq(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqxacc_QVbVb(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqxacc_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqb_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.eq(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eq_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eq_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.eq(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqand_QVhVh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqand_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqh_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.eq(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqor_QVhVh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqor_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqh_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.eq(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqxacc_QVhVh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqxacc_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqh_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.eq(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eq_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eq_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.eq(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqand_QVwVw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqand_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqw_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.eq(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqor_QVwVw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqor_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqw_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.eq(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqxacc_QVwVw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqxacc_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqw_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.gt(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gt_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gt_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.gt(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtand_QVbVb(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtand_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtb_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.gt(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtor_QVbVb(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtor_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtb_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.gt(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtxacc_QVbVb(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtxacc_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtb_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.gt(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gt_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gt_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgth) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.gt(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtand_QVhVh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtand_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgth_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.gt(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtor_QVhVh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtor_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgth_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.gt(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtxacc_QVhVh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtxacc_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgth_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.gt(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gt_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gt_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.gt(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtand_QVubVub(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtand_QVubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtub_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.gt(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtor_QVubVub(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtor_QVubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtub_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.gt(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtxacc_QVubVub(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtxacc_QVubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtub_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.gt(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gt_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gt_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.gt(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtand_QVuhVuh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtand_QVuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuh_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.gt(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtor_QVuhVuh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtor_QVuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuh_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.gt(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtxacc_QVuhVuh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtxacc_QVuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuh_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.gt(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gt_VuwVuw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gt_VuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.gt(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtand_QVuwVuw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtand_QVuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuw_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.gt(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtor_QVuwVuw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtor_QVuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuw_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.gt(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtxacc_QVuwVuw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtxacc_QVuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuw_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.gt(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gt_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gt_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.gt(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtand_QVwVw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtand_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtw_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.gt(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtor_QVwVw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtor_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtw_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.gt(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtxacc_QVwVw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtxacc_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtw_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w=vinsert(Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vinsert_VwR(HVX_Vector Vx, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vinsert_VwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vinsertwr) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vlalign(Vu32,Vv32,Rt8) + C Intrinsic Prototype: HVX_Vector Q6_V_vlalign_VVR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vlalign_VVR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlalignb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vlalign(Vu32,Vv32,#u3) + C Intrinsic Prototype: HVX_Vector Q6_V_vlalign_VVI(HVX_Vector Vu, HVX_Vector Vv, Word32 Iu3) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vlalign_VVI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlalignbi) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vlsr(Vu32.uh,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vlsr_VuhR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vlsr_VuhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlsrh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vlsr(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vlsr_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vlsr_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlsrhv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vlsr(Vu32.uw,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vlsr_VuwR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vlsr_VuwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlsrw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vlsr(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vlsr_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vlsr_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlsrwv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vlut32(Vu32.b,Vv32.b,Rt8) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vlut32_VbVbR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vlut32_VbVbR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvvb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.b|=vlut32(Vu32.b,Vv32.b,Rt8) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vlut32or_VbVbVbR(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vlut32or_VbVbVbR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvvb_oracc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vlut16(Vu32.b,Vv32.h,Rt8) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vlut16_VbVhR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vlut16_VbVhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvwh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h|=vlut16(Vu32.b,Vv32.h,Rt8) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vlut16or_WhVbVhR(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vlut16or_WhVbVhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvwh_oracc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmax(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmax_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vmax_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaxh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vmax(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vub_vmax_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vmax_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaxub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vmax(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vmax_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vmax_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaxuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmax(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmax_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vmax_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaxw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmin(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmin_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vmin_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vminh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vmin(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vub_vmin_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vmin_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vminub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vmin(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vmin_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vmin_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vminuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmin(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmin_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vmin_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vminw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpa(Vuu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpa_WubRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpa_WubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpabus) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vmpa(Vuu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpaacc_WhWubRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpaacc_WhWubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpabus_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpa(Vuu32.ub,Vvv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpa_WubWb(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpa_WubWb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpabusv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpa(Vuu32.ub,Vvv32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpa_WubWub(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpa_WubWub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpabuuv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vmpa(Vuu32.h,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpa_WhRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpa_WhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpahb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vmpa(Vuu32.h,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpaacc_WwWhRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpaacc_WwWhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpahb_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpy(Vu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpy_VubRb(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpy_VubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpybus) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vmpy(Vu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpyacc_WhVubRb(HVX_VectorPair Vxx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpyacc_WhVubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpybus_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpy(Vu32.ub,Vv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpy_VubVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpy_VubVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpybusv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vmpy(Vu32.ub,Vv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpyacc_WhVubVb(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpyacc_WhVubVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpybusv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpy(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpy_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpy_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpybv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vmpy(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpyacc_WhVbVb(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpyacc_WhVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpybv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpye(Vu32.w,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpye_VwVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpye_VwVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyewuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vmpy(Vu32.h,Rt32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpy_VhRh(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpy_VhRh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vmpy(Vu32.h,Rt32.h):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpyacc_WwVhRh_sat(HVX_VectorPair Vxx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpyacc_WwVhRh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhsat_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmpy(Vu32.h,Rt32.h):<<1:rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpy_VhRh_s1_rnd_sat(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpy_VhRh_s1_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhsrs) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmpy(Vu32.h,Rt32.h):<<1:sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpy_VhRh_s1_sat(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpy_VhRh_s1_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhss) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vmpy(Vu32.h,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpy_VhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpy_VhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhus) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vmpy(Vu32.h,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpyacc_WwVhVuh(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpyacc_WwVhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhus_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vmpy(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpy_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpy_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vmpy(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpyacc_WwVhVh(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpyacc_WwVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmpy(Vu32.h,Vv32.h):<<1:rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpy_VhVh_s1_rnd_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpy_VhVh_s1_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhvsrs) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyieo(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyieo_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyieo_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyieoh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyie(Vu32.w,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyieacc_VwVwVh(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyieacc_VwVwVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiewh_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyie(Vu32.w,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyie_VwVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyie_VwVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiewuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyie(Vu32.w,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyieacc_VwVwVuh(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyieacc_VwVwVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiewuh_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmpyi(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpyi_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpyi_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyih) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.h+=vmpyi(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpyiacc_VhVhVh(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpyiacc_VhVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyih_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmpyi(Vu32.h,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpyi_VhRb(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpyi_VhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyihb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.h+=vmpyi(Vu32.h,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpyiacc_VhVhRb(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpyiacc_VhVhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyihb_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyio(Vu32.w,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyio_VwVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyio_VwVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiowh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyi(Vu32.w,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyi_VwRb(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyi_VwRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiwb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyi(Vu32.w,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyiacc_VwVwRb(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyiacc_VwVwRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiwb_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyi(Vu32.w,Rt32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyi_VwRh(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyi_VwRh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiwh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyi(Vu32.w,Rt32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyiacc_VwVwRh(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyiacc_VwVwRh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiwh_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyo(Vu32.w,Vv32.h):<<1:sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyo_VwVh_s1_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyo_VwVh_s1_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyowh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyo(Vu32.w,Vv32.h):<<1:rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyo_VwVh_s1_rnd_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyo_VwVh_s1_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyowh_rnd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyo(Vu32.w,Vv32.h):<<1:rnd:sat:shift + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyoacc_VwVwVh_s1_rnd_sat_shift(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyoacc_VwVwVh_s1_rnd_sat_shift __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyowh_rnd_sacc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyo(Vu32.w,Vv32.h):<<1:sat:shift + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyoacc_VwVwVh_s1_sat_shift(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyoacc_VwVwVh_s1_sat_shift __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyowh_sacc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uh=vmpy(Vu32.ub,Rt32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vmpy_VubRub(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuh_vmpy_VubRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uh+=vmpy(Vu32.ub,Rt32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vmpyacc_WuhVubRub(HVX_VectorPair Vxx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuh_vmpyacc_WuhVubRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyub_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uh=vmpy(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vmpy_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuh_vmpy_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyubv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uh+=vmpy(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vmpyacc_WuhVubVub(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuh_vmpyacc_WuhVubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyubv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vmpy(Vu32.uh,Rt32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vmpy_VuhRuh(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vmpy_VuhRuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uw+=vmpy(Vu32.uh,Rt32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vmpyacc_WuwVuhRuh(HVX_VectorPair Vxx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vmpyacc_WuwVuhRuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyuh_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vmpy(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vmpy_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vmpy_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyuhv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uw+=vmpy(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vmpyacc_WuwVuhVuh(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vmpyacc_WuwVuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyuhv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vmux(Qt4,Vu32,Vv32) + C Intrinsic Prototype: HVX_Vector Q6_V_vmux_QVV(HVX_VectorPred Qt, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vmux_QVV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmux) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vnavg(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vnavg_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vnavg_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnavgh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vnavg(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vnavg_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vnavg_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnavgub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vnavg(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vnavg_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vnavg_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnavgw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vnormamt(Vu32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vnormamt_Vh(HVX_Vector Vu) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vnormamt_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnormamth) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vnormamt(Vu32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vnormamt_Vw(HVX_Vector Vu) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vnormamt_Vw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnormamtw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vnot(Vu32) + C Intrinsic Prototype: HVX_Vector Q6_V_vnot_V(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vnot_V __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnot) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vor(Vu32,Vv32) + C Intrinsic Prototype: HVX_Vector Q6_V_vor_VV(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vor_VV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vpacke(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vpacke_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vpacke_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackeb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vpacke(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vpacke_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vpacke_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackeh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vpack(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vpack_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vpack_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackhb_sat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vpack(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vpack_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vpack_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackhub_sat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vpacko(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vpacko_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vpacko_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackob) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vpacko(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vpacko_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vpacko_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackoh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vpack(Vu32.w,Vv32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vpack_VwVw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vpack_VwVw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackwh_sat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vpack(Vu32.w,Vv32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vpack_VwVw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vpack_VwVw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackwuh_sat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vpopcount(Vu32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vpopcount_Vh(HVX_Vector Vu) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vpopcount_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpopcounth) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vrdelta(Vu32,Vv32) + C Intrinsic Prototype: HVX_Vector Q6_V_vrdelta_VV(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vrdelta_VV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrdelta) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vrmpy(Vu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vrmpy_VubRb(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vrmpy_VubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybus) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vrmpy(Vu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vrmpyacc_VwVubRb(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vrmpyacc_VwVubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybus_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vrmpy(Vuu32.ub,Rt32.b,#u1) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vrmpy_WubRbI(HVX_VectorPair Vuu, Word32 Rt, Word32 Iu1) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vrmpy_WubRbI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybusi) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vrmpy(Vuu32.ub,Rt32.b,#u1) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vrmpyacc_WwWubRbI(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt, Word32 Iu1) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vrmpyacc_WwWubRbI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybusi_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vrmpy(Vu32.ub,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vrmpy_VubVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vrmpy_VubVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybusv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vrmpy(Vu32.ub,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vrmpyacc_VwVubVb(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vrmpyacc_VwVubVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybusv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vrmpy(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vrmpy_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vrmpy_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vrmpy(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vrmpyacc_VwVbVb(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vrmpyacc_VwVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vrmpy(Vu32.ub,Rt32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vrmpy_VubRub(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vrmpy_VubRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpyub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.uw+=vrmpy(Vu32.ub,Rt32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vrmpyacc_VuwVubRub(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vrmpyacc_VuwVubRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpyub_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vrmpy(Vuu32.ub,Rt32.ub,#u1) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vrmpy_WubRubI(HVX_VectorPair Vuu, Word32 Rt, Word32 Iu1) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vrmpy_WubRubI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpyubi) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uw+=vrmpy(Vuu32.ub,Rt32.ub,#u1) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vrmpyacc_WuwWubRubI(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt, Word32 Iu1) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vrmpyacc_WuwWubRubI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpyubi_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vrmpy(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vrmpy_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vrmpy_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpyubv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.uw+=vrmpy(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vrmpyacc_VuwVubVub(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vrmpyacc_VuwVubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpyubv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vror(Vu32,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_V_vror_VR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vror_VR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vror) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vround(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vround_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vround_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vroundhb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vround(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vround_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vround_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vroundhub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vround(Vu32.w,Vv32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vround_VwVw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vround_VwVw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vroundwh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vround(Vu32.w,Vv32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vround_VwVw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vround_VwVw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vroundwuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vrsad(Vuu32.ub,Rt32.ub,#u1) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vrsad_WubRubI(HVX_VectorPair Vuu, Word32 Rt, Word32 Iu1) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vrsad_WubRubI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrsadubi) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uw+=vrsad(Vuu32.ub,Rt32.ub,#u1) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vrsadacc_WuwWubRubI(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt, Word32 Iu1) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vrsadacc_WuwWubRubI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrsadubi_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vsat(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vub_vsat_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vsat_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsathub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vsat(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vsat_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vsat_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsatwh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vsxt(Vu32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vsxt_Vb(HVX_Vector Vu) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vsxt_Vb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vsxt(Vu32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vsxt_Vh(HVX_Vector Vu) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vsxt_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vshuffe(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vshuffe_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vshuffe_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshufeh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vshuff(Vu32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vshuff_Vb(HVX_Vector Vu) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vshuff_Vb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshuffb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vshuffe(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vshuffe_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vshuffe_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshuffeb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vshuff(Vu32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vshuff_Vh(HVX_Vector Vu) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vshuff_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshuffh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vshuffo(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vshuffo_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vshuffo_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshuffob) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32=vshuff(Vu32,Vv32,Rt8) + C Intrinsic Prototype: HVX_VectorPair Q6_W_vshuff_VVR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_W_vshuff_VVR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshuffvdd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.b=vshuffoe(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wb_vshuffoe_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wb_vshuffoe_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshufoeb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vshuffoe(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vshuffoe_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vshuffoe_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshufoeh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vshuffo(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vshuffo_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vshuffo_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshufoh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vsub(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vsub_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vsub_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.b=vsub(Vuu32.b,Vvv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wb_vsub_WbWb(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wb_vsub_WbWb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubb_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) Vx32.b-=Vu32.b + C Intrinsic Prototype: HVX_Vector Q6_Vb_condnac_QnVbVb(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_condnac_QnVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubbnq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) Vx32.b-=Vu32.b + C Intrinsic Prototype: HVX_Vector Q6_Vb_condnac_QVbVb(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_condnac_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubbq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vsub(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vsub_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vsub_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vsub(Vuu32.h,Vvv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vsub_WhWh(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vsub_WhWh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubh_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) Vx32.h-=Vu32.h + C Intrinsic Prototype: HVX_Vector Q6_Vh_condnac_QnVhVh(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_condnac_QnVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubhnq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) Vx32.h-=Vu32.h + C Intrinsic Prototype: HVX_Vector Q6_Vh_condnac_QVhVh(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_condnac_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubhq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vsub(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vsub_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vsub_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vsub(Vuu32.h,Vvv32.h):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vsub_WhWh_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vsub_WhWh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubhsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vsub(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vsub_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vsub_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubhw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vsub(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vsub_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vsub_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsububh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vsub(Vu32.ub,Vv32.ub):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vsub_VubVub_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vsub_VubVub_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsububsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.ub=vsub(Vuu32.ub,Vvv32.ub):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wub_vsub_WubWub_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wub_vsub_WubWub_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsububsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vsub(Vu32.uh,Vv32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vsub_VuhVuh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vsub_VuhVuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubuhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uh=vsub(Vuu32.uh,Vvv32.uh):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vsub_WuhWuh_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuh_vsub_WuhWuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubuhsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vsub(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vsub_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vsub_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubuhw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vsub(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vsub_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vsub_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vsub(Vuu32.w,Vvv32.w) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vsub_WwWw(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vsub_WwWw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubw_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) Vx32.w-=Vu32.w + C Intrinsic Prototype: HVX_Vector Q6_Vw_condnac_QnVwVw(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_condnac_QnVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubwnq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) Vx32.w-=Vu32.w + C Intrinsic Prototype: HVX_Vector Q6_Vw_condnac_QVwVw(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_condnac_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubwq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vsub(Vu32.w,Vv32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vsub_VwVw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vsub_VwVw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubwsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vsub(Vuu32.w,Vvv32.w):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vsub_WwWw_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vsub_WwWw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubwsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32=vswap(Qt4,Vu32,Vv32) + C Intrinsic Prototype: HVX_VectorPair Q6_W_vswap_QVV(HVX_VectorPred Qt, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_W_vswap_QVV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vswap) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vtmpy(Vuu32.b,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vtmpy_WbRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vtmpy_WbRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vtmpyb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vtmpy(Vuu32.b,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vtmpyacc_WhWbRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vtmpyacc_WhWbRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vtmpyb_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vtmpy(Vuu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vtmpy_WubRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vtmpy_WubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vtmpybus) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vtmpy(Vuu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vtmpyacc_WhWubRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vtmpyacc_WhWubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vtmpybus_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vtmpy(Vuu32.h,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vtmpy_WhRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vtmpy_WhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vtmpyhb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vtmpy(Vuu32.h,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vtmpyacc_WwWhRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vtmpyacc_WwWhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vtmpyhb_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vunpack(Vu32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vunpack_Vb(HVX_Vector Vu) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vunpack_Vb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vunpackb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vunpack(Vu32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vunpack_Vh(HVX_Vector Vu) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vunpack_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vunpackh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h|=vunpacko(Vu32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vunpackoor_WhVb(HVX_VectorPair Vxx, HVX_Vector Vu) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vunpackoor_WhVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vunpackob) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w|=vunpacko(Vu32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vunpackoor_WwVh(HVX_VectorPair Vxx, HVX_Vector Vu) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vunpackoor_WwVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vunpackoh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uh=vunpack(Vu32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vunpack_Vub(HVX_Vector Vu) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuh_vunpack_Vub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vunpackub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vunpack(Vu32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vunpack_Vuh(HVX_Vector Vu) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuw_vunpack_Vuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vunpackuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vxor(Vu32,Vv32) + C Intrinsic Prototype: HVX_Vector Q6_V_vxor_VV(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vxor_VV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vxor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uh=vzxt(Vu32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vzxt_Vub(HVX_Vector Vu) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuh_vzxt_Vub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vzb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vzxt(Vu32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vzxt_Vuh(HVX_Vector Vu) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuw_vzxt_Vuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vzh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vsplat(Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vsplat_R(Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vb_vsplat_R __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_lvsplatb) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.h=vsplat(Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vsplat_R(Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vsplat_R __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_lvsplath) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Qd4=vsetq2(Rt32) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vsetq2_R(Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vsetq2_R __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_scalar2v2) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Qd4.b=vshuffe(Qs4.h,Qt4.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Qb_vshuffe_QhQh(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Qb_vshuffe_QhQh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_shuffeqh) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Qd4.h=vshuffe(Qs4.w,Qt4.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Qh_vshuffe_QwQw(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Qh_vshuffe_QwQw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_shuffeqw) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vadd(Vu32.b,Vv32.b):sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vadd_VbVb_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vadd_VbVb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddbsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.b=vadd(Vuu32.b,Vvv32.b):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wb_vadd_WbWb_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wb_vadd_WbWb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddbsat_dv) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.w=vadd(Vu32.w,Vv32.w,Qx4):carry + C Intrinsic Prototype: HVX_Vector Q6_Vw_vadd_VwVwQ_carry(HVX_Vector Vu, HVX_Vector Vv, HVX_VectorPred* Qx) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vadd_VwVwQ_carry __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddcarry) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.h=vadd(vclb(Vu32.h),Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vadd_vclb_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vadd_vclb_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddclbh) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.w=vadd(vclb(Vu32.w),Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vadd_vclb_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vadd_vclb_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddclbw) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vadd(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vaddacc_WwVhVh(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vaddacc_WwVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddhw_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vadd(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vaddacc_WhVubVub(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vaddacc_WhVubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddubh_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vadd(Vu32.ub,Vv32.b):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vadd_VubVb_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vadd_VubVb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddububb_sat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vadd(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vaddacc_WwVuhVuh(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vaddacc_WwVuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vadduhw_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vadd(Vu32.uw,Vv32.uw):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vadd_VuwVuw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vadd_VuwVuw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vadduwsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vadd(Vuu32.uw,Vvv32.uw):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vadd_WuwWuw_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuw_vadd_WuwWuw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vadduwsat_dv) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32=vand(!Qu4,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_V_vand_QnR(HVX_VectorPred Qu, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_V_vand_QnR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandnqrt) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vx32|=vand(!Qu4,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_V_vandor_VQnR(HVX_Vector Vx, HVX_VectorPred Qu, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_V_vandor_VQnR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandnqrt_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32=vand(!Qv4,Vu32) + C Intrinsic Prototype: HVX_Vector Q6_V_vand_QnV(HVX_VectorPred Qv, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vand_QnV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandvnqv) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32=vand(Qv4,Vu32) + C Intrinsic Prototype: HVX_Vector Q6_V_vand_QV(HVX_VectorPred Qv, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vand_QV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandvqv) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vasr(Vu32.h,Vv32.h,Rt8):sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vasr_VhVhR_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vasr_VhVhR_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrhbsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vasr(Vu32.uw,Vv32.uw,Rt8):rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vasr_VuwVuwR_rnd_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vasr_VuwVuwR_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasruwuhrndsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vasr(Vu32.w,Vv32.w,Rt8):rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vasr_VwVwR_rnd_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vasr_VwVwR_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrwuhrndsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vlsr(Vu32.ub,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vub_vlsr_VubR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vlsr_VubR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlsrb) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vlut32(Vu32.b,Vv32.b,Rt8):nomatch + C Intrinsic Prototype: HVX_Vector Q6_Vb_vlut32_VbVbR_nomatch(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vlut32_VbVbR_nomatch __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvvb_nm) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vx32.b|=vlut32(Vu32.b,Vv32.b,#u3) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vlut32or_VbVbVbI(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv, Word32 Iu3) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vlut32or_VbVbVbI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvvb_oracci) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vlut32(Vu32.b,Vv32.b,#u3) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vlut32_VbVbI(HVX_Vector Vu, HVX_Vector Vv, Word32 Iu3) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vlut32_VbVbI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvvbi) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vlut16(Vu32.b,Vv32.h,Rt8):nomatch + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vlut16_VbVhR_nomatch(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vlut16_VbVhR_nomatch __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvwh_nm) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vxx32.h|=vlut16(Vu32.b,Vv32.h,#u3) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vlut16or_WhVbVhI(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv, Word32 Iu3) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vlut16or_WhVbVhI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvwh_oracci) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vlut16(Vu32.b,Vv32.h,#u3) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vlut16_VbVhI(HVX_Vector Vu, HVX_Vector Vv, Word32 Iu3) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vlut16_VbVhI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvwhi) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vmax(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vmax_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vmax_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaxb) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vmin(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vmin_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vmin_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vminb) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vmpa(Vuu32.uh,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpa_WuhRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpa_WuhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpauhb) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vmpa(Vuu32.uh,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpaacc_WwWuhRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpaacc_WwWuhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpauhb_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32=vmpye(Vu32.w,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_W_vmpye_VwVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_W_vmpye_VwVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyewuh_64) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyi(Vu32.w,Rt32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyi_VwRub(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyi_VwRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiwub) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyi(Vu32.w,Rt32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyiacc_VwVwRub(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyiacc_VwVwRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiwub_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vxx32+=vmpyo(Vu32.w,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_W_vmpyoacc_WVwVh(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_W_vmpyoacc_WVwVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyowh_64_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vround(Vu32.uh,Vv32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vround_VuhVuh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vround_VuhVuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrounduhub) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vround(Vu32.uw,Vv32.uw):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vround_VuwVuw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vround_VuwVuw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrounduwuh) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vsat(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vsat_VuwVuw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vsat_VuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsatuwuh) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vsub(Vu32.b,Vv32.b):sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vsub_VbVb_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vsub_VbVb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubbsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.b=vsub(Vuu32.b,Vvv32.b):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wb_vsub_WbWb_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wb_vsub_WbWb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubbsat_dv) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.w=vsub(Vu32.w,Vv32.w,Qx4):carry + C Intrinsic Prototype: HVX_Vector Q6_Vw_vsub_VwVwQ_carry(HVX_Vector Vu, HVX_Vector Vv, HVX_VectorPred* Qx) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vsub_VwVwQ_carry __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubcarry) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vsub(Vu32.ub,Vv32.b):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vsub_VubVb_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vsub_VubVb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubububb_sat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vsub(Vu32.uw,Vv32.uw):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vsub_VuwVuw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vsub_VuwVuw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubuwsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vsub(Vuu32.uw,Vvv32.uw):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vsub_WuwWuw_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuw_vsub_WuwWuw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubuwsat_dv) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.b=vabs(Vu32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vabs_Vb(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vabs_Vb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsb) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.b=vabs(Vu32.b):sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vabs_Vb_sat(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vabs_Vb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsb_sat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vx32.h+=vasl(Vu32.h,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vaslacc_VhVhR(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vaslacc_VhVhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaslh_acc) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vx32.h+=vasr(Vu32.h,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasracc_VhVhR(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasracc_VhVhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrh_acc) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vasr(Vu32.uh,Vv32.uh,Rt8):rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vasr_VuhVuhR_rnd_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vasr_VuhVuhR_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasruhubrndsat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vasr(Vu32.uh,Vv32.uh,Rt8):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vasr_VuhVuhR_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vasr_VuhVuhR_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasruhubsat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vasr(Vu32.uw,Vv32.uw,Rt8):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vasr_VuwVuwR_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vasr_VuwVuwR_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasruwuhsat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.b=vavg(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vavg_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vavg_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgb) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.b=vavg(Vu32.b,Vv32.b):rnd + C Intrinsic Prototype: HVX_Vector Q6_Vb_vavg_VbVb_rnd(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vavg_VbVb_rnd __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgbrnd) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vavg(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vavg_VuwVuw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vavg_VuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavguw) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vavg(Vu32.uw,Vv32.uw):rnd + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vavg_VuwVuw_rnd(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vavg_VuwVuw_rnd __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavguwrnd) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vdd32=#0 + C Intrinsic Prototype: HVX_VectorPair Q6_W_vzero() + Instruction Type: MAPPING + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_W_vzero __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdd0) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vtmp.h=vgather(Rt32,Mu2,Vv32.h).h + C Intrinsic Prototype: void Q6_vgather_ARMVh(HVX_Vector* Rs, Word32 Rt, Word32 Mu, HVX_Vector Vv) + Instruction Type: CVI_GATHER + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_vgather_ARMVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgathermh) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: if (Qs4) vtmp.h=vgather(Rt32,Mu2,Vv32.h).h + C Intrinsic Prototype: void Q6_vgather_AQRMVh(HVX_Vector* Rs, HVX_VectorPred Qs, Word32 Rt, Word32 Mu, HVX_Vector Vv) + Instruction Type: CVI_GATHER + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_vgather_AQRMVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgathermhq) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vtmp.h=vgather(Rt32,Mu2,Vvv32.w).h + C Intrinsic Prototype: void Q6_vgather_ARMWw(HVX_Vector* Rs, Word32 Rt, Word32 Mu, HVX_VectorPair Vvv) + Instruction Type: CVI_GATHER_DV + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_vgather_ARMWw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgathermhw) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: if (Qs4) vtmp.h=vgather(Rt32,Mu2,Vvv32.w).h + C Intrinsic Prototype: void Q6_vgather_AQRMWw(HVX_Vector* Rs, HVX_VectorPred Qs, Word32 Rt, Word32 Mu, HVX_VectorPair Vvv) + Instruction Type: CVI_GATHER_DV + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_vgather_AQRMWw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgathermhwq) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vtmp.w=vgather(Rt32,Mu2,Vv32.w).w + C Intrinsic Prototype: void Q6_vgather_ARMVw(HVX_Vector* Rs, Word32 Rt, Word32 Mu, HVX_Vector Vv) + Instruction Type: CVI_GATHER + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_vgather_ARMVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgathermw) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: if (Qs4) vtmp.w=vgather(Rt32,Mu2,Vv32.w).w + C Intrinsic Prototype: void Q6_vgather_AQRMVw(HVX_Vector* Rs, HVX_VectorPred Qs, Word32 Rt, Word32 Mu, HVX_Vector Vv) + Instruction Type: CVI_GATHER + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_vgather_AQRMVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgathermwq) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.h=vlut4(Vu32.uh,Rtt32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vlut4_VuhPh(HVX_Vector Vu, Word64 Rtt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT2 + ========================================================================== */ + +#define Q6_Vh_vlut4_VuhPh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlut4) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpa(Vuu32.ub,Rt32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpa_WubRub(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpa_WubRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpabuu) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vmpa(Vuu32.ub,Rt32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpaacc_WhWubRub(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpaacc_WhWubRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpabuu_acc) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vx32.h=vmpa(Vx32.h,Vu32.h,Rtt32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpa_VhVhVhPh_sat(HVX_Vector Vx, HVX_Vector Vu, Word64 Rtt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT2 + ========================================================================== */ + +#define Q6_Vh_vmpa_VhVhVhPh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpahhsat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vx32.h=vmpa(Vx32.h,Vu32.uh,Rtt32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpa_VhVhVuhPuh_sat(HVX_Vector Vx, HVX_Vector Vu, Word64 Rtt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT2 + ========================================================================== */ + +#define Q6_Vh_vmpa_VhVhVuhPuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpauhuhsat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vx32.h=vmps(Vx32.h,Vu32.uh,Rtt32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmps_VhVhVuhPuh_sat(HVX_Vector Vx, HVX_Vector Vu, Word64 Rtt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT2 + ========================================================================== */ + +#define Q6_Vh_vmps_VhVhVuhPuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpsuhuhsat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vmpy(Vu32.h,Rt32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpyacc_WwVhRh(HVX_VectorPair Vxx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpyacc_WwVhRh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyh_acc) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vmpye(Vu32.uh,Rt32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vmpye_VuhRuh(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vmpye_VuhRuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyuhe) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vx32.uw+=vmpye(Vu32.uh,Rt32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vmpyeacc_VuwVuhRuh(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vmpyeacc_VuwVuhRuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyuhe_acc) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.b=vnavg(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vnavg_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vnavg_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnavgb) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.b=prefixsum(Qv4) + C Intrinsic Prototype: HVX_Vector Q6_Vb_prefixsum_Q(HVX_VectorPred Qv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_prefixsum_Q __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vprefixqb) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.h=prefixsum(Qv4) + C Intrinsic Prototype: HVX_Vector Q6_Vh_prefixsum_Q(HVX_VectorPred Qv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_prefixsum_Q __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vprefixqh) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.w=prefixsum(Qv4) + C Intrinsic Prototype: HVX_Vector Q6_Vw_prefixsum_Q(HVX_VectorPred Qv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_prefixsum_Q __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vprefixqw) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vscatter(Rt32,Mu2,Vv32.h).h=Vw32 + C Intrinsic Prototype: void Q6_vscatter_RMVhV(Word32 Rt, Word32 Mu, HVX_Vector Vv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatter_RMVhV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermh) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vscatter(Rt32,Mu2,Vv32.h).h+=Vw32 + C Intrinsic Prototype: void Q6_vscatteracc_RMVhV(Word32 Rt, Word32 Mu, HVX_Vector Vv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatteracc_RMVhV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermh_add) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: if (Qs4) vscatter(Rt32,Mu2,Vv32.h).h=Vw32 + C Intrinsic Prototype: void Q6_vscatter_QRMVhV(HVX_VectorPred Qs, Word32 Rt, Word32 Mu, HVX_Vector Vv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatter_QRMVhV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermhq) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vscatter(Rt32,Mu2,Vvv32.w).h=Vw32 + C Intrinsic Prototype: void Q6_vscatter_RMWwV(Word32 Rt, Word32 Mu, HVX_VectorPair Vvv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER_DV + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatter_RMWwV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermhw) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vscatter(Rt32,Mu2,Vvv32.w).h+=Vw32 + C Intrinsic Prototype: void Q6_vscatteracc_RMWwV(Word32 Rt, Word32 Mu, HVX_VectorPair Vvv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER_DV + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatteracc_RMWwV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermhw_add) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: if (Qs4) vscatter(Rt32,Mu2,Vvv32.w).h=Vw32 + C Intrinsic Prototype: void Q6_vscatter_QRMWwV(HVX_VectorPred Qs, Word32 Rt, Word32 Mu, HVX_VectorPair Vvv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER_DV + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatter_QRMWwV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermhwq) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vscatter(Rt32,Mu2,Vv32.w).w=Vw32 + C Intrinsic Prototype: void Q6_vscatter_RMVwV(Word32 Rt, Word32 Mu, HVX_Vector Vv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatter_RMVwV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermw) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vscatter(Rt32,Mu2,Vv32.w).w+=Vw32 + C Intrinsic Prototype: void Q6_vscatteracc_RMVwV(Word32 Rt, Word32 Mu, HVX_Vector Vv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatteracc_RMVwV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermw_add) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: if (Qs4) vscatter(Rt32,Mu2,Vv32.w).w=Vw32 + C Intrinsic Prototype: void Q6_vscatter_QRMVwV(HVX_VectorPred Qs, Word32 Rt, Word32 Mu, HVX_Vector Vv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatter_QRMVwV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermwq) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Vd32.w=vadd(Vu32.w,Vv32.w,Qs4):carry:sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vadd_VwVwQ_carry_sat(HVX_Vector Vu, HVX_Vector Vv, HVX_VectorPred Qs) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vadd_VwVwQ_carry_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddcarrysat) +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HVX_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Vxx32.w=vasrinto(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vasrinto_WwVwVw(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vasrinto_WwVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasr_into) +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HVX_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vrotr(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vrotr_VuwVuw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vrotr_VuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrotr) +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HVX_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Vd32.w=vsatdw(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vsatdw_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vsatdw_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsatdw) +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HVX_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Vdd32.w=v6mpy(Vuu32.ub,Vvv32.b,#u2):h + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_v6mpy_WubWbI_h(HVX_VectorPair Vuu, HVX_VectorPair Vvv, Word32 Iu2) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_v6mpy_WubWbI_h __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_v6mpyhubs10) +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HVX_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=v6mpy(Vuu32.ub,Vvv32.b,#u2):h + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_v6mpyacc_WwWubWbI_h(HVX_VectorPair Vxx, HVX_VectorPair Vuu, HVX_VectorPair Vvv, Word32 Iu2) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_v6mpyacc_WwWubWbI_h __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_v6mpyhubs10_vxx) +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HVX_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Vdd32.w=v6mpy(Vuu32.ub,Vvv32.b,#u2):v + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_v6mpy_WubWbI_v(HVX_VectorPair Vuu, HVX_VectorPair Vvv, Word32 Iu2) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_v6mpy_WubWbI_v __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_v6mpyvubs10) +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HVX_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=v6mpy(Vuu32.ub,Vvv32.b,#u2):v + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_v6mpyacc_WwWubWbI_v(HVX_VectorPair Vxx, HVX_VectorPair Vuu, HVX_VectorPair Vvv, Word32 Iu2) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_v6mpyacc_WwWubWbI_v __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_v6mpyvubs10_vxx) +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#endif /* __HVX__ */ + +#endif diff --git a/lib/include/immintrin.h b/lib/include/immintrin.h index 22f7a520c9..56d3dadf6a 100644 --- a/lib/include/immintrin.h +++ b/lib/include/immintrin.h @@ -72,11 +72,6 @@ #include #endif -#if !(defined(_MSC_VER) || defined(__SCE__)) || __has_feature(modules) || \ - defined(__VPCLMULQDQ__) -#include -#endif - /* No feature check desired due to internal checks */ #include @@ -230,6 +225,11 @@ #include #endif +#if !(defined(_MSC_VER) || defined(__SCE__)) || __has_feature(modules) || \ + defined(__VPCLMULQDQ__) +#include +#endif + #if !(defined(_MSC_VER) || defined(__SCE__)) || __has_feature(modules) || \ defined(__VAES__) #include diff --git a/lib/include/intrin.h b/lib/include/intrin.h index a78b96997d..ff8eb8fca2 100644 --- a/lib/include/intrin.h +++ b/lib/include/intrin.h @@ -451,24 +451,47 @@ unsigned char _InterlockedCompareExchange128_rel(__int64 volatile *_Destination, static __inline__ void __DEFAULT_FN_ATTRS __movsb(unsigned char *__dst, unsigned char const *__src, size_t __n) { - __asm__ __volatile__("rep movsb" : "+D"(__dst), "+S"(__src), "+c"(__n) - : : "memory"); +#if defined(__x86_64__) + __asm__ __volatile__("rep movsb" + : "+D"(__dst), "+S"(__src), "+c"(__n) + : + : "memory"); +#else + __asm__ __volatile__("xchg %%esi, %1\nrep movsb\nxchg %%esi, %1" + : "+D"(__dst), "+r"(__src), "+c"(__n) + : + : "memory"); +#endif } static __inline__ void __DEFAULT_FN_ATTRS __movsd(unsigned long *__dst, unsigned long const *__src, size_t __n) { +#if defined(__x86_64__) __asm__ __volatile__("rep movsl" : "+D"(__dst), "+S"(__src), "+c"(__n) : : "memory"); +#else + __asm__ __volatile__("xchg %%esi, %1\nrep movsl\nxchg %%esi, %1" + : "+D"(__dst), "+r"(__src), "+c"(__n) + : + : "memory"); +#endif } static __inline__ void __DEFAULT_FN_ATTRS __movsw(unsigned short *__dst, unsigned short const *__src, size_t __n) { +#if defined(__x86_64__) __asm__ __volatile__("rep movsw" : "+D"(__dst), "+S"(__src), "+c"(__n) : : "memory"); +#else + __asm__ __volatile__("xchg %%esi, %1\nrep movsw\nxchg %%esi, %1" + : "+D"(__dst), "+r"(__src), "+c"(__n) + : + : "memory"); +#endif } static __inline__ void __DEFAULT_FN_ATTRS __stosd(unsigned long *__dst, unsigned long __x, @@ -507,16 +530,26 @@ static __inline__ void __DEFAULT_FN_ATTRS __stosq(unsigned __int64 *__dst, |* Misc \*----------------------------------------------------------------------------*/ #if defined(__i386__) || defined(__x86_64__) +#if defined(__i386__) +#define __cpuid_count(__leaf, __count, __eax, __ebx, __ecx, __edx) \ + __asm("cpuid" \ + : "=a"(__eax), "=b"(__ebx), "=c"(__ecx), "=d"(__edx) \ + : "0"(__leaf), "2"(__count)) +#else +/* x86-64 uses %rbx as the base register, so preserve it. */ +#define __cpuid_count(__leaf, __count, __eax, __ebx, __ecx, __edx) \ + __asm("xchgq %%rbx,%q1\n" \ + "cpuid\n" \ + "xchgq %%rbx,%q1" \ + : "=a"(__eax), "=r"(__ebx), "=c"(__ecx), "=d"(__edx) \ + : "0"(__leaf), "2"(__count)) +#endif static __inline__ void __DEFAULT_FN_ATTRS __cpuid(int __info[4], int __level) { - __asm__("cpuid" - : "=a"(__info[0]), "=b"(__info[1]), "=c"(__info[2]), "=d"(__info[3]) - : "a"(__level), "c"(0)); + __cpuid_count(__level, 0, __info[0], __info[1], __info[2], __info[3]); } static __inline__ void __DEFAULT_FN_ATTRS __cpuidex(int __info[4], int __level, int __ecx) { - __asm__("cpuid" - : "=a"(__info[0]), "=b"(__info[1]), "=c"(__info[2]), "=d"(__info[3]) - : "a"(__level), "c"(__ecx)); + __cpuid_count(__level, __ecx, __info[0], __info[1], __info[2], __info[3]); } static __inline__ void __DEFAULT_FN_ATTRS __halt(void) { __asm__ volatile("hlt"); diff --git a/lib/include/keylockerintrin.h b/lib/include/keylockerintrin.h index c15d39c8e3..68b0a56896 100644 --- a/lib/include/keylockerintrin.h +++ b/lib/include/keylockerintrin.h @@ -230,10 +230,12 @@ _mm_aesenc128kl_u8(__m128i* __odata, __m128i __idata, const void *__h) { /// HandleKeyType (Handle[511:0]) != HANDLE_KEY_TYPE_AES256 ) /// IF (IllegalHandle) /// ZF := 1 +/// MEM[__odata+127:__odata] := 0 /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate512 (Handle[511:0], IWKey) /// IF (Authentic == 0) /// ZF := 1 +/// MEM[__odata+127:__odata] := 0 /// ELSE /// MEM[__odata+127:__odata] := AES256Encrypt (__idata[127:0], UnwrappedKey) /// ZF := 0 @@ -267,10 +269,12 @@ _mm_aesenc256kl_u8(__m128i* __odata, __m128i __idata, const void *__h) { /// HandleKeyType (Handle[383:0]) != HANDLE_KEY_TYPE_AES128) /// IF (IllegalHandle) /// ZF := 1 +/// MEM[__odata+127:__odata] := 0 /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate384 (Handle[383:0], IWKey) /// IF (Authentic == 0) /// ZF := 1 +/// MEM[__odata+127:__odata] := 0 /// ELSE /// MEM[__odata+127:__odata] := AES128Decrypt (__idata[127:0], UnwrappedKey) /// ZF := 0 @@ -304,10 +308,12 @@ _mm_aesdec128kl_u8(__m128i* __odata, __m128i __idata, const void *__h) { /// HandleKeyType (Handle[511:0]) != HANDLE_KEY_TYPE_AES256) /// IF (IllegalHandle) /// ZF := 1 +/// MEM[__odata+127:__odata] := 0 /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate512 (Handle[511:0], IWKey) /// IF (Authentic == 0) /// ZF := 1 +/// MEM[__odata+127:__odata] := 0 /// ELSE /// MEM[__odata+127:__odata] := AES256Decrypt (__idata[127:0], UnwrappedKey) /// ZF := 0 @@ -354,10 +360,16 @@ _mm_aesdec256kl_u8(__m128i* __odata, __m128i __idata, const void *__h) { /// HandleKeyType (Handle[383:0]) != HANDLE_KEY_TYPE_AES128 ) /// IF (IllegalHandle) /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate384 (Handle[383:0], IWKey) /// IF Authentic == 0 /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// FOR i := 0 to 7 /// __odata[i] := AES128Encrypt (__idata[i], UnwrappedKey) @@ -394,10 +406,16 @@ _mm_aesencwide128kl_u8(__m128i __odata[8], const __m128i __idata[8], const void* /// HandleKeyType (Handle[511:0]) != HANDLE_KEY_TYPE_AES512 ) /// IF (IllegalHandle) /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate512 (Handle[511:0], IWKey) /// IF Authentic == 0 /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// FOR i := 0 to 7 /// __odata[i] := AES256Encrypt (__idata[i], UnwrappedKey) @@ -434,10 +452,16 @@ _mm_aesencwide256kl_u8(__m128i __odata[8], const __m128i __idata[8], const void* /// HandleKeyType (Handle) != HANDLE_KEY_TYPE_AES128 ) /// IF (IllegalHandle) /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate384 (Handle[383:0], IWKey) /// IF Authentic == 0 /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// FOR i := 0 to 7 /// __odata[i] := AES128Decrypt (__idata[i], UnwrappedKey) @@ -474,10 +498,16 @@ _mm_aesdecwide128kl_u8(__m128i __odata[8], const __m128i __idata[8], const void* /// HandleKeyType (Handle) != HANDLE_KEY_TYPE_AES512 ) /// If (IllegalHandle) /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate512 (Handle[511:0], IWKey) /// IF Authentic == 0 /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// FOR i := 0 to 7 /// __odata[i] := AES256Decrypt (__idata[i], UnwrappedKey) diff --git a/lib/include/opencl-c-base.h b/lib/include/opencl-c-base.h index e8dcd70377..3c5e2c9739 100644 --- a/lib/include/opencl-c-base.h +++ b/lib/include/opencl-c-base.h @@ -21,9 +21,37 @@ #define cl_khr_subgroup_shuffle 1 #define cl_khr_subgroup_shuffle_relative 1 #define cl_khr_subgroup_clustered_reduce 1 +#define cl_khr_extended_bit_ops 1 +#define cl_khr_integer_dot_product 1 +#define __opencl_c_integer_dot_product_input_4x8bit 1 +#define __opencl_c_integer_dot_product_input_4x8bit_packed 1 + #endif // defined(__SPIR__) #endif // (defined(__OPENCL_CPP_VERSION__) || __OPENCL_C_VERSION__ >= 200) +// Define feature macros for OpenCL C 2.0 +#if (defined(__OPENCL_CPP_VERSION__) || __OPENCL_C_VERSION__ == 200) +#define __opencl_c_pipes 1 +#define __opencl_c_generic_address_space 1 +#define __opencl_c_work_group_collective_functions 1 +#define __opencl_c_atomic_order_acq_rel 1 +#define __opencl_c_atomic_order_seq_cst 1 +#define __opencl_c_atomic_scope_device 1 +#define __opencl_c_atomic_scope_all_devices 1 +#define __opencl_c_device_enqueue 1 +#define __opencl_c_read_write_images 1 +#define __opencl_c_program_scope_global_variables 1 +#define __opencl_c_images 1 +#endif + +// Define header-only feature macros for OpenCL C 3.0. +#if (__OPENCL_C_VERSION__ == 300) +// For the SPIR target all features are supported. +#if defined(__SPIR__) +#define __opencl_c_atomic_scope_all_devices 1 +#endif // defined(__SPIR__) +#endif // (__OPENCL_C_VERSION__ == 300) + // built-in scalar data types: /** @@ -141,7 +169,9 @@ typedef double double8 __attribute__((ext_vector_type(8))); typedef double double16 __attribute__((ext_vector_type(16))); #endif -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__OPENCL_CPP_VERSION__) +#define NULL nullptr +#elif defined(__OPENCL_C_VERSION__) #define NULL ((void*)0) #endif @@ -297,7 +327,12 @@ typedef enum memory_scope { memory_scope_work_item = __OPENCL_MEMORY_SCOPE_WORK_ITEM, memory_scope_work_group = __OPENCL_MEMORY_SCOPE_WORK_GROUP, memory_scope_device = __OPENCL_MEMORY_SCOPE_DEVICE, +#if defined(__opencl_c_atomic_scope_all_devices) memory_scope_all_svm_devices = __OPENCL_MEMORY_SCOPE_ALL_SVM_DEVICES, +#if (__OPENCL_C_VERSION__ >= CL_VERSION_3_0) + memory_scope_all_devices = memory_scope_all_svm_devices, +#endif // __OPENCL_C_VERSION__ >= CL_VERSION_3_0 +#endif // defined(__opencl_c_atomic_scope_all_devices) #if defined(cl_intel_subgroups) || defined(cl_khr_subgroups) memory_scope_sub_group = __OPENCL_MEMORY_SCOPE_SUB_GROUP #endif @@ -322,7 +357,9 @@ typedef enum memory_order memory_order_acquire = __ATOMIC_ACQUIRE, memory_order_release = __ATOMIC_RELEASE, memory_order_acq_rel = __ATOMIC_ACQ_REL, +#if defined(__opencl_c_atomic_order_seq_cst) memory_order_seq_cst = __ATOMIC_SEQ_CST +#endif } memory_order; #endif // defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) @@ -445,8 +482,113 @@ typedef struct { #endif // defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +/** + * OpenCL v1.1/1.2/2.0 s6.2.4.2 - as_type operators + * Reinterprets a data type as another data type of the same size + */ +#define as_char(x) __builtin_astype((x), char) +#define as_char2(x) __builtin_astype((x), char2) +#define as_char3(x) __builtin_astype((x), char3) +#define as_char4(x) __builtin_astype((x), char4) +#define as_char8(x) __builtin_astype((x), char8) +#define as_char16(x) __builtin_astype((x), char16) + +#define as_uchar(x) __builtin_astype((x), uchar) +#define as_uchar2(x) __builtin_astype((x), uchar2) +#define as_uchar3(x) __builtin_astype((x), uchar3) +#define as_uchar4(x) __builtin_astype((x), uchar4) +#define as_uchar8(x) __builtin_astype((x), uchar8) +#define as_uchar16(x) __builtin_astype((x), uchar16) + +#define as_short(x) __builtin_astype((x), short) +#define as_short2(x) __builtin_astype((x), short2) +#define as_short3(x) __builtin_astype((x), short3) +#define as_short4(x) __builtin_astype((x), short4) +#define as_short8(x) __builtin_astype((x), short8) +#define as_short16(x) __builtin_astype((x), short16) + +#define as_ushort(x) __builtin_astype((x), ushort) +#define as_ushort2(x) __builtin_astype((x), ushort2) +#define as_ushort3(x) __builtin_astype((x), ushort3) +#define as_ushort4(x) __builtin_astype((x), ushort4) +#define as_ushort8(x) __builtin_astype((x), ushort8) +#define as_ushort16(x) __builtin_astype((x), ushort16) + +#define as_int(x) __builtin_astype((x), int) +#define as_int2(x) __builtin_astype((x), int2) +#define as_int3(x) __builtin_astype((x), int3) +#define as_int4(x) __builtin_astype((x), int4) +#define as_int8(x) __builtin_astype((x), int8) +#define as_int16(x) __builtin_astype((x), int16) + +#define as_uint(x) __builtin_astype((x), uint) +#define as_uint2(x) __builtin_astype((x), uint2) +#define as_uint3(x) __builtin_astype((x), uint3) +#define as_uint4(x) __builtin_astype((x), uint4) +#define as_uint8(x) __builtin_astype((x), uint8) +#define as_uint16(x) __builtin_astype((x), uint16) + +#define as_long(x) __builtin_astype((x), long) +#define as_long2(x) __builtin_astype((x), long2) +#define as_long3(x) __builtin_astype((x), long3) +#define as_long4(x) __builtin_astype((x), long4) +#define as_long8(x) __builtin_astype((x), long8) +#define as_long16(x) __builtin_astype((x), long16) + +#define as_ulong(x) __builtin_astype((x), ulong) +#define as_ulong2(x) __builtin_astype((x), ulong2) +#define as_ulong3(x) __builtin_astype((x), ulong3) +#define as_ulong4(x) __builtin_astype((x), ulong4) +#define as_ulong8(x) __builtin_astype((x), ulong8) +#define as_ulong16(x) __builtin_astype((x), ulong16) + +#define as_float(x) __builtin_astype((x), float) +#define as_float2(x) __builtin_astype((x), float2) +#define as_float3(x) __builtin_astype((x), float3) +#define as_float4(x) __builtin_astype((x), float4) +#define as_float8(x) __builtin_astype((x), float8) +#define as_float16(x) __builtin_astype((x), float16) + +#ifdef cl_khr_fp64 +#define as_double(x) __builtin_astype((x), double) +#define as_double2(x) __builtin_astype((x), double2) +#define as_double3(x) __builtin_astype((x), double3) +#define as_double4(x) __builtin_astype((x), double4) +#define as_double8(x) __builtin_astype((x), double8) +#define as_double16(x) __builtin_astype((x), double16) +#endif // cl_khr_fp64 + +#ifdef cl_khr_fp16 +#define as_half(x) __builtin_astype((x), half) +#define as_half2(x) __builtin_astype((x), half2) +#define as_half3(x) __builtin_astype((x), half3) +#define as_half4(x) __builtin_astype((x), half4) +#define as_half8(x) __builtin_astype((x), half8) +#define as_half16(x) __builtin_astype((x), half16) +#endif // cl_khr_fp16 + +#define as_size_t(x) __builtin_astype((x), size_t) +#define as_ptrdiff_t(x) __builtin_astype((x), ptrdiff_t) +#define as_intptr_t(x) __builtin_astype((x), intptr_t) +#define as_uintptr_t(x) __builtin_astype((x), uintptr_t) + +// OpenCL v1.1 s6.9, v1.2/2.0 s6.10 - Function qualifiers + +#define __kernel_exec(X, typen) __kernel \ + __attribute__((work_group_size_hint(X, 1, 1))) \ + __attribute__((vec_type_hint(typen))) + +#define kernel_exec(X, typen) __kernel \ + __attribute__((work_group_size_hint(X, 1, 1))) \ + __attribute__((vec_type_hint(typen))) + +#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2) +// OpenCL v1.2 s6.12.13, v2.0 s6.13.13 - printf + +int printf(__constant const char* st, ...) __attribute__((format(printf, 1, 2))); +#endif + #ifdef cl_intel_device_side_avc_motion_estimation -#pragma OPENCL EXTENSION cl_intel_device_side_avc_motion_estimation : begin #define CLK_AVC_ME_MAJOR_16x16_INTEL 0x0 #define CLK_AVC_ME_MAJOR_16x8_INTEL 0x1 @@ -580,7 +722,6 @@ typedef struct { #define CLK_AVC_IME_RESULT_DUAL_REFERENCE_STREAMOUT_INITIALIZE_INTEL 0x0 #define CLK_AVC_IME_RESULT_DUAL_REFERENCE_STREAMIN_INITIALIZE_INTEL 0x0 -#pragma OPENCL EXTENSION cl_intel_device_side_avc_motion_estimation : end #endif // cl_intel_device_side_avc_motion_estimation // Disable any extensions we may have enabled previously. diff --git a/lib/include/opencl-c.h b/lib/include/opencl-c.h index ab665628c8..fc50dd718c 100644 --- a/lib/include/opencl-c.h +++ b/lib/include/opencl-c.h @@ -23,10 +23,11 @@ #endif //cl_khr_3d_image_writes #endif //__OPENCL_C_VERSION__ < CL_VERSION_2_0 -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2) + +#if (defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2)) && defined(__SPIR__) #pragma OPENCL EXTENSION cl_intel_planar_yuv : begin #pragma OPENCL EXTENSION cl_intel_planar_yuv : end -#endif // defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2) +#endif // (defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2)) && defined(__SPIR__) #define __ovld __attribute__((overloadable)) #define __conv __attribute__((convergent)) @@ -6339,101 +6340,6 @@ half16 __ovld __cnfn convert_half16_rtz(double16); #endif // cl_khr_fp16 -/** - * OpenCL v1.1/1.2/2.0 s6.2.4.2 - as_type operators - * Reinterprets a data type as another data type of the same size - */ -#define as_char(x) __builtin_astype((x), char) -#define as_char2(x) __builtin_astype((x), char2) -#define as_char3(x) __builtin_astype((x), char3) -#define as_char4(x) __builtin_astype((x), char4) -#define as_char8(x) __builtin_astype((x), char8) -#define as_char16(x) __builtin_astype((x), char16) - -#define as_uchar(x) __builtin_astype((x), uchar) -#define as_uchar2(x) __builtin_astype((x), uchar2) -#define as_uchar3(x) __builtin_astype((x), uchar3) -#define as_uchar4(x) __builtin_astype((x), uchar4) -#define as_uchar8(x) __builtin_astype((x), uchar8) -#define as_uchar16(x) __builtin_astype((x), uchar16) - -#define as_short(x) __builtin_astype((x), short) -#define as_short2(x) __builtin_astype((x), short2) -#define as_short3(x) __builtin_astype((x), short3) -#define as_short4(x) __builtin_astype((x), short4) -#define as_short8(x) __builtin_astype((x), short8) -#define as_short16(x) __builtin_astype((x), short16) - -#define as_ushort(x) __builtin_astype((x), ushort) -#define as_ushort2(x) __builtin_astype((x), ushort2) -#define as_ushort3(x) __builtin_astype((x), ushort3) -#define as_ushort4(x) __builtin_astype((x), ushort4) -#define as_ushort8(x) __builtin_astype((x), ushort8) -#define as_ushort16(x) __builtin_astype((x), ushort16) - -#define as_int(x) __builtin_astype((x), int) -#define as_int2(x) __builtin_astype((x), int2) -#define as_int3(x) __builtin_astype((x), int3) -#define as_int4(x) __builtin_astype((x), int4) -#define as_int8(x) __builtin_astype((x), int8) -#define as_int16(x) __builtin_astype((x), int16) - -#define as_uint(x) __builtin_astype((x), uint) -#define as_uint2(x) __builtin_astype((x), uint2) -#define as_uint3(x) __builtin_astype((x), uint3) -#define as_uint4(x) __builtin_astype((x), uint4) -#define as_uint8(x) __builtin_astype((x), uint8) -#define as_uint16(x) __builtin_astype((x), uint16) - -#define as_long(x) __builtin_astype((x), long) -#define as_long2(x) __builtin_astype((x), long2) -#define as_long3(x) __builtin_astype((x), long3) -#define as_long4(x) __builtin_astype((x), long4) -#define as_long8(x) __builtin_astype((x), long8) -#define as_long16(x) __builtin_astype((x), long16) - -#define as_ulong(x) __builtin_astype((x), ulong) -#define as_ulong2(x) __builtin_astype((x), ulong2) -#define as_ulong3(x) __builtin_astype((x), ulong3) -#define as_ulong4(x) __builtin_astype((x), ulong4) -#define as_ulong8(x) __builtin_astype((x), ulong8) -#define as_ulong16(x) __builtin_astype((x), ulong16) - -#define as_float(x) __builtin_astype((x), float) -#define as_float2(x) __builtin_astype((x), float2) -#define as_float3(x) __builtin_astype((x), float3) -#define as_float4(x) __builtin_astype((x), float4) -#define as_float8(x) __builtin_astype((x), float8) -#define as_float16(x) __builtin_astype((x), float16) - -#ifdef cl_khr_fp64 -#define as_double(x) __builtin_astype((x), double) -#define as_double2(x) __builtin_astype((x), double2) -#define as_double3(x) __builtin_astype((x), double3) -#define as_double4(x) __builtin_astype((x), double4) -#define as_double8(x) __builtin_astype((x), double8) -#define as_double16(x) __builtin_astype((x), double16) -#endif //cl_khr_fp64 - -#ifdef cl_khr_fp16 -#define as_half(x) __builtin_astype((x), half) -#define as_half2(x) __builtin_astype((x), half2) -#define as_half3(x) __builtin_astype((x), half3) -#define as_half4(x) __builtin_astype((x), half4) -#define as_half8(x) __builtin_astype((x), half8) -#define as_half16(x) __builtin_astype((x), half16) -#endif //cl_khr_fp16 - -// OpenCL v1.1 s6.9, v1.2/2.0 s6.10 - Function qualifiers - -#define __kernel_exec(X, typen) __kernel \ - __attribute__((work_group_size_hint(X, 1, 1))) \ - __attribute__((vec_type_hint(typen))) - -#define kernel_exec(X, typen) __kernel \ - __attribute__((work_group_size_hint(X, 1, 1))) \ - __attribute__((vec_type_hint(typen))) - // OpenCL v1.1 s6.11.1, v1.2 s6.12.1, v2.0 s6.13.1 - Work-item Functions /** @@ -6494,8 +6400,7 @@ size_t __ovld __cnfn get_local_id(uint dimindx); * Returns the number of work-groups that will execute a * kernel for dimension identified by dimindx. * Valid values of dimindx are 0 to get_work_dim() - 1. - * For other values of dimindx, get_num_groups () returns - * 1. + * For other values of dimindx, get_num_groups() returns 1. * For clEnqueueTask, this always returns 1. */ size_t __ovld __cnfn get_num_groups(uint dimindx); @@ -7354,7 +7259,7 @@ half16 __ovld __cnfn fmod(half16 x, half16 y); * Returns fmin(x - floor (x), 0x1.fffffep-1f ). * floor(x) is returned in iptr. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld fract(float x, float *iptr); float2 __ovld fract(float2 x, float2 *iptr); float3 __ovld fract(float3 x, float3 *iptr); @@ -7436,7 +7341,7 @@ half4 __ovld fract(half4 x, __private half4 *iptr); half8 __ovld fract(half8 x, __private half8 *iptr); half16 __ovld fract(half16 x, __private half16 *iptr); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Extract mantissa and exponent from x. For each @@ -7444,7 +7349,7 @@ half16 __ovld fract(half16 x, __private half16 *iptr); * magnitude in the interval [1/2, 1) or 0. Each * component of x equals mantissa returned * 2^exp. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld frexp(float x, int *exp); float2 __ovld frexp(float2 x, int2 *exp); float3 __ovld frexp(float3 x, int3 *exp); @@ -7526,7 +7431,7 @@ half4 __ovld frexp(half4 x, __private int4 *exp); half8 __ovld frexp(half8 x, __private int8 *exp); half16 __ovld frexp(half16 x, __private int16 *exp); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Compute the value of the square root of x^2 + y^2 @@ -7651,7 +7556,7 @@ half8 __ovld __cnfn lgamma(half8 x); half16 __ovld __cnfn lgamma(half16 x); #endif //cl_khr_fp16 -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld lgamma_r(float x, int *signp); float2 __ovld lgamma_r(float2 x, int2 *signp); float3 __ovld lgamma_r(float3 x, int3 *signp); @@ -7733,7 +7638,7 @@ half4 __ovld lgamma_r(half4 x, __private int4 *signp); half8 __ovld lgamma_r(half8 x, __private int8 *signp); half16 __ovld lgamma_r(half16 x, __private int16 *signp); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Compute natural logarithm. @@ -7957,7 +7862,7 @@ half16 __ovld __cnfn minmag(half16 x, half16 y); * the argument. It stores the integral part in the object * pointed to by iptr. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld modf(float x, float *iptr); float2 __ovld modf(float2 x, float2 *iptr); float3 __ovld modf(float3 x, float3 *iptr); @@ -8039,7 +7944,7 @@ half4 __ovld modf(half4 x, __private half4 *iptr); half8 __ovld modf(half8 x, __private half8 *iptr); half16 __ovld modf(half16 x, __private half16 *iptr); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Returns a quiet NaN. The nancode may be placed @@ -8217,7 +8122,7 @@ half16 __ovld __cnfn remainder(half16 x, half16 y); * sign as x/y. It stores this signed value in the object * pointed to by quo. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld remquo(float x, float y, int *quo); float2 __ovld remquo(float2 x, float2 y, int2 *quo); float3 __ovld remquo(float3 x, float3 y, int3 *quo); @@ -8300,7 +8205,7 @@ half4 __ovld remquo(half4 x, half4 y, __private int4 *quo); half8 __ovld remquo(half8 x, half8 y, __private int8 *quo); half16 __ovld remquo(half16 x, half16 y, __private int16 *quo); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Round to integral value (using round to nearest * even rounding mode) in floating-point format. @@ -8441,7 +8346,7 @@ half16 __ovld __cnfn sin(half16); * is the return value and computed cosine is returned * in cosval. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld sincos(float x, float *cosval); float2 __ovld sincos(float2 x, float2 *cosval); float3 __ovld sincos(float3 x, float3 *cosval); @@ -8523,7 +8428,7 @@ half4 __ovld sincos(half4 x, __private half4 *cosval); half8 __ovld sincos(half8 x, __private half8 *cosval); half16 __ovld sincos(half16 x, __private half16 *cosval); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Compute hyperbolic sine. @@ -9449,54 +9354,54 @@ ulong16 __ovld __cnfn clz(ulong16 x); * component type of x, if x is a vector. */ #if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) -char __ovld ctz(char x); -uchar __ovld ctz(uchar x); -char2 __ovld ctz(char2 x); -uchar2 __ovld ctz(uchar2 x); -char3 __ovld ctz(char3 x); -uchar3 __ovld ctz(uchar3 x); -char4 __ovld ctz(char4 x); -uchar4 __ovld ctz(uchar4 x); -char8 __ovld ctz(char8 x); -uchar8 __ovld ctz(uchar8 x); -char16 __ovld ctz(char16 x); -uchar16 __ovld ctz(uchar16 x); -short __ovld ctz(short x); -ushort __ovld ctz(ushort x); -short2 __ovld ctz(short2 x); -ushort2 __ovld ctz(ushort2 x); -short3 __ovld ctz(short3 x); -ushort3 __ovld ctz(ushort3 x); -short4 __ovld ctz(short4 x); -ushort4 __ovld ctz(ushort4 x); -short8 __ovld ctz(short8 x); -ushort8 __ovld ctz(ushort8 x); -short16 __ovld ctz(short16 x); -ushort16 __ovld ctz(ushort16 x); -int __ovld ctz(int x); -uint __ovld ctz(uint x); -int2 __ovld ctz(int2 x); -uint2 __ovld ctz(uint2 x); -int3 __ovld ctz(int3 x); -uint3 __ovld ctz(uint3 x); -int4 __ovld ctz(int4 x); -uint4 __ovld ctz(uint4 x); -int8 __ovld ctz(int8 x); -uint8 __ovld ctz(uint8 x); -int16 __ovld ctz(int16 x); -uint16 __ovld ctz(uint16 x); -long __ovld ctz(long x); -ulong __ovld ctz(ulong x); -long2 __ovld ctz(long2 x); -ulong2 __ovld ctz(ulong2 x); -long3 __ovld ctz(long3 x); -ulong3 __ovld ctz(ulong3 x); -long4 __ovld ctz(long4 x); -ulong4 __ovld ctz(ulong4 x); -long8 __ovld ctz(long8 x); -ulong8 __ovld ctz(ulong8 x); -long16 __ovld ctz(long16 x); -ulong16 __ovld ctz(ulong16 x); +char __ovld __cnfn ctz(char x); +uchar __ovld __cnfn ctz(uchar x); +char2 __ovld __cnfn ctz(char2 x); +uchar2 __ovld __cnfn ctz(uchar2 x); +char3 __ovld __cnfn ctz(char3 x); +uchar3 __ovld __cnfn ctz(uchar3 x); +char4 __ovld __cnfn ctz(char4 x); +uchar4 __ovld __cnfn ctz(uchar4 x); +char8 __ovld __cnfn ctz(char8 x); +uchar8 __ovld __cnfn ctz(uchar8 x); +char16 __ovld __cnfn ctz(char16 x); +uchar16 __ovld __cnfn ctz(uchar16 x); +short __ovld __cnfn ctz(short x); +ushort __ovld __cnfn ctz(ushort x); +short2 __ovld __cnfn ctz(short2 x); +ushort2 __ovld __cnfn ctz(ushort2 x); +short3 __ovld __cnfn ctz(short3 x); +ushort3 __ovld __cnfn ctz(ushort3 x); +short4 __ovld __cnfn ctz(short4 x); +ushort4 __ovld __cnfn ctz(ushort4 x); +short8 __ovld __cnfn ctz(short8 x); +ushort8 __ovld __cnfn ctz(ushort8 x); +short16 __ovld __cnfn ctz(short16 x); +ushort16 __ovld __cnfn ctz(ushort16 x); +int __ovld __cnfn ctz(int x); +uint __ovld __cnfn ctz(uint x); +int2 __ovld __cnfn ctz(int2 x); +uint2 __ovld __cnfn ctz(uint2 x); +int3 __ovld __cnfn ctz(int3 x); +uint3 __ovld __cnfn ctz(uint3 x); +int4 __ovld __cnfn ctz(int4 x); +uint4 __ovld __cnfn ctz(uint4 x); +int8 __ovld __cnfn ctz(int8 x); +uint8 __ovld __cnfn ctz(uint8 x); +int16 __ovld __cnfn ctz(int16 x); +uint16 __ovld __cnfn ctz(uint16 x); +long __ovld __cnfn ctz(long x); +ulong __ovld __cnfn ctz(ulong x); +long2 __ovld __cnfn ctz(long2 x); +ulong2 __ovld __cnfn ctz(ulong2 x); +long3 __ovld __cnfn ctz(long3 x); +ulong3 __ovld __cnfn ctz(ulong3 x); +long4 __ovld __cnfn ctz(long4 x); +ulong4 __ovld __cnfn ctz(ulong4 x); +long8 __ovld __cnfn ctz(long8 x); +ulong8 __ovld __cnfn ctz(ulong8 x); +long16 __ovld __cnfn ctz(long16 x); +ulong16 __ovld __cnfn ctz(ulong16 x); #endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) /** @@ -10002,6 +9907,7 @@ ulong16 __ovld __cnfn upsample(uint16 hi, uint16 lo); /* * popcount(x): returns the number of set bit in x */ +#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2) char __ovld __cnfn popcount(char x); uchar __ovld __cnfn popcount(uchar x); char2 __ovld __cnfn popcount(char2 x); @@ -10050,6 +9956,7 @@ long8 __ovld __cnfn popcount(long8 x); ulong8 __ovld __cnfn popcount(ulong8 x); long16 __ovld __cnfn popcount(long16 x); ulong16 __ovld __cnfn popcount(ulong16 x); +#endif // defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2) /** * Multiply two 24-bit integer values x and y and add @@ -11342,7 +11249,7 @@ half8 __ovld vload8(size_t offset, const __constant half *p); half16 __ovld vload16(size_t offset, const __constant half *p); #endif //cl_khr_fp16 -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) char2 __ovld vload2(size_t offset, const char *p); uchar2 __ovld vload2(size_t offset, const uchar *p); short2 __ovld vload2(size_t offset, const short *p); @@ -11580,9 +11487,9 @@ half4 __ovld vload4(size_t offset, const __private half *p); half8 __ovld vload8(size_t offset, const __private half *p); half16 __ovld vload16(size_t offset, const __private half *p); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) void __ovld vstore2(char2 data, size_t offset, char *p); void __ovld vstore2(uchar2 data, size_t offset, uchar *p); void __ovld vstore2(short2 data, size_t offset, short *p); @@ -11816,7 +11723,7 @@ void __ovld vstore4(half4 data, size_t offset, __private half *p); void __ovld vstore8(half8 data, size_t offset, __private half *p); void __ovld vstore16(half16 data, size_t offset, __private half *p); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Read sizeof (half) bytes of data from address @@ -11827,13 +11734,13 @@ void __ovld vstore16(half16 data, size_t offset, __private half *p); * must be 16-bit aligned. */ float __ovld vload_half(size_t offset, const __constant half *p); -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld vload_half(size_t offset, const half *p); #else float __ovld vload_half(size_t offset, const __global half *p); float __ovld vload_half(size_t offset, const __local half *p); float __ovld vload_half(size_t offset, const __private half *p); -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Read sizeof (halfn) bytes of data from address @@ -11848,7 +11755,7 @@ float3 __ovld vload_half3(size_t offset, const __constant half *p); float4 __ovld vload_half4(size_t offset, const __constant half *p); float8 __ovld vload_half8(size_t offset, const __constant half *p); float16 __ovld vload_half16(size_t offset, const __constant half *p); -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float2 __ovld vload_half2(size_t offset, const half *p); float3 __ovld vload_half3(size_t offset, const half *p); float4 __ovld vload_half4(size_t offset, const half *p); @@ -11870,7 +11777,7 @@ float3 __ovld vload_half3(size_t offset, const __private half *p); float4 __ovld vload_half4(size_t offset, const __private half *p); float8 __ovld vload_half8(size_t offset, const __private half *p); float16 __ovld vload_half16(size_t offset, const __private half *p); -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * The float value given by data is first @@ -11883,7 +11790,7 @@ float16 __ovld vload_half16(size_t offset, const __private half *p); * The default current rounding mode is round to * nearest even. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) void __ovld vstore_half(float data, size_t offset, half *p); void __ovld vstore_half_rte(float data, size_t offset, half *p); void __ovld vstore_half_rtz(float data, size_t offset, half *p); @@ -11929,7 +11836,7 @@ void __ovld vstore_half_rtz(double data, size_t offset, __private half *p); void __ovld vstore_half_rtp(double data, size_t offset, __private half *p); void __ovld vstore_half_rtn(double data, size_t offset, __private half *p); #endif //cl_khr_fp64 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * The floatn value given by data is converted to @@ -11942,7 +11849,7 @@ void __ovld vstore_half_rtn(double data, size_t offset, __private half *p); * The default current rounding mode is round to * nearest even. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) void __ovld vstore_half2(float2 data, size_t offset, half *p); void __ovld vstore_half3(float3 data, size_t offset, half *p); void __ovld vstore_half4(float4 data, size_t offset, half *p); @@ -12148,7 +12055,7 @@ void __ovld vstore_half4_rtn(double4 data, size_t offset, __private half *p); void __ovld vstore_half8_rtn(double8 data, size_t offset, __private half *p); void __ovld vstore_half16_rtn(double16 data, size_t offset, __private half *p); #endif //cl_khr_fp64 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * For n = 1, 2, 4, 8 and 16 read sizeof (halfn) @@ -12169,7 +12076,7 @@ float3 __ovld vloada_half3(size_t offset, const __constant half *p); float4 __ovld vloada_half4(size_t offset, const __constant half *p); float8 __ovld vloada_half8(size_t offset, const __constant half *p); float16 __ovld vloada_half16(size_t offset, const __constant half *p); -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld vloada_half(size_t offset, const half *p); float2 __ovld vloada_half2(size_t offset, const half *p); float3 __ovld vloada_half3(size_t offset, const half *p); @@ -12195,7 +12102,7 @@ float3 __ovld vloada_half3(size_t offset, const __private half *p); float4 __ovld vloada_half4(size_t offset, const __private half *p); float8 __ovld vloada_half8(size_t offset, const __private half *p); float16 __ovld vloada_half16(size_t offset, const __private half *p); -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * The floatn value given by data is converted to @@ -12213,7 +12120,7 @@ float16 __ovld vloada_half16(size_t offset, const __private half *p); * mode. The default current rounding mode is * round to nearest even. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) void __ovld vstorea_half(float data, size_t offset, half *p); void __ovld vstorea_half2(float2 data, size_t offset, half *p); void __ovld vstorea_half3(float3 data, size_t offset, half *p); @@ -12498,7 +12405,7 @@ void __ovld vstorea_half4_rtn(double4 data,size_t offset, __private half *p); void __ovld vstorea_half8_rtn(double8 data,size_t offset, __private half *p); void __ovld vstorea_half16_rtn(double16 data,size_t offset, __private half *p); #endif //cl_khr_fp64 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) // OpenCL v1.1 s6.11.8, v1.2 s6.12.8, v2.0 s6.13.8 - Synchronization Functions @@ -12582,7 +12489,7 @@ void __ovld write_mem_fence(cl_mem_fence_flags flags); // OpenCL v2.0 s6.13.9 - Address Space Qualifier Functions -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) cl_mem_fence_flags __ovld get_fence(const void *ptr); cl_mem_fence_flags __ovld get_fence(void *ptr); @@ -12593,7 +12500,7 @@ cl_mem_fence_flags __ovld get_fence(void *ptr); * where gentype is builtin type or user defined type. */ -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) // OpenCL v1.1 s6.11.10, v1.2 s6.12.10, v2.0 s6.13.10 - Async Copies from Global to Local Memory, Local to Global Memory, and Prefetch @@ -13397,291 +13304,324 @@ void __ovld atomic_init(volatile atomic_double *object, double value); void __ovld atomic_work_item_fence(cl_mem_fence_flags flags, memory_order order, memory_scope scope); // atomic_fetch() +// OpenCL v2.0 s6.13.11.7.5: +// add/sub: atomic type argument can be uintptr_t/intptr_t, value type argument can be ptrdiff_t. +#if defined(__opencl_c_atomic_order_seq_cst) && defined(__opencl_c_atomic_scope_device) int __ovld atomic_fetch_add(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_add_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_add_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_add(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_add_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_add_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); int __ovld atomic_fetch_sub(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_sub_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_sub_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_sub(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_sub_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_sub_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); int __ovld atomic_fetch_or(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_or_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_or_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_or(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_or_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_or_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); int __ovld atomic_fetch_xor(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_xor_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_xor_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_xor(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_xor_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_xor_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); int __ovld atomic_fetch_and(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_and_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_and_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_and(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_and_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_and_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); int __ovld atomic_fetch_min(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_min_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_min_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_min(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_min_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_min_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); int __ovld atomic_fetch_max(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_max_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_max_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_max(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_max_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_max_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); #if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) long __ovld atomic_fetch_add(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_add_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_add_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_add(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_add_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_add_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); long __ovld atomic_fetch_sub(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_sub_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_sub_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_sub(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_sub_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_sub_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); long __ovld atomic_fetch_or(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_or_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_or_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_or(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_or_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_or_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); long __ovld atomic_fetch_xor(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_xor_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_xor_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_xor(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_xor_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_xor_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); long __ovld atomic_fetch_and(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_and_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_and_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_and(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_and_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_and_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); long __ovld atomic_fetch_min(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_min_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_min_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_min(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_min_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_min_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); long __ovld atomic_fetch_max(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_max_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_max_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_max(volatile atomic_ulong *object, ulong operand); +uintptr_t __ovld atomic_fetch_add(volatile atomic_uintptr_t *object, ptrdiff_t operand); +uintptr_t __ovld atomic_fetch_sub(volatile atomic_uintptr_t *object, ptrdiff_t operand); +#endif //defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#endif + +#if defined(__opencl_c_atomic_scope_device) +int __ovld atomic_fetch_add_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_add_explicit(volatile atomic_uint *object, uint operand, memory_order order); +int __ovld atomic_fetch_sub_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_sub_explicit(volatile atomic_uint *object, uint operand, memory_order order); +int __ovld atomic_fetch_or_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_or_explicit(volatile atomic_uint *object, uint operand, memory_order order); +int __ovld atomic_fetch_xor_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_xor_explicit(volatile atomic_uint *object, uint operand, memory_order order); +int __ovld atomic_fetch_and_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_and_explicit(volatile atomic_uint *object, uint operand, memory_order order); +int __ovld atomic_fetch_min_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_min_explicit(volatile atomic_uint *object, uint operand, memory_order order); +int __ovld atomic_fetch_max_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_max_explicit(volatile atomic_uint *object, uint operand, memory_order order); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +long __ovld atomic_fetch_add_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_add_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +long __ovld atomic_fetch_sub_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_sub_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +long __ovld atomic_fetch_or_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_or_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +long __ovld atomic_fetch_xor_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_xor_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +long __ovld atomic_fetch_and_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_and_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +long __ovld atomic_fetch_min_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_min_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +long __ovld atomic_fetch_max_explicit(volatile atomic_long *object, long operand, memory_order order); ulong __ovld atomic_fetch_max_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +uintptr_t __ovld atomic_fetch_add_explicit(volatile atomic_uintptr_t *object, ptrdiff_t operand, memory_order order); +uintptr_t __ovld atomic_fetch_sub_explicit(volatile atomic_uintptr_t *object, ptrdiff_t operand, memory_order order); +#endif //defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#endif + +int __ovld atomic_fetch_add_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_add_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +int __ovld atomic_fetch_sub_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_sub_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +int __ovld atomic_fetch_or_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_or_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +int __ovld atomic_fetch_xor_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_xor_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +int __ovld atomic_fetch_and_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_and_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +int __ovld atomic_fetch_min_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_min_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +int __ovld atomic_fetch_max_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_max_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +long __ovld atomic_fetch_add_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_add_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +long __ovld atomic_fetch_sub_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_sub_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +long __ovld atomic_fetch_or_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_or_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +long __ovld atomic_fetch_xor_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_xor_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +long __ovld atomic_fetch_and_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_and_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +long __ovld atomic_fetch_min_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_min_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +long __ovld atomic_fetch_max_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_max_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); #endif //defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) - -// OpenCL v2.0 s6.13.11.7.5: -// add/sub: atomic type argument can be uintptr_t/intptr_t, value type argument can be ptrdiff_t. -// or/xor/and/min/max: atomic type argument can be intptr_t/uintptr_t, value type argument can be intptr_t/uintptr_t. - #if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) -uintptr_t __ovld atomic_fetch_add(volatile atomic_uintptr_t *object, ptrdiff_t operand); -uintptr_t __ovld atomic_fetch_add_explicit(volatile atomic_uintptr_t *object, ptrdiff_t operand, memory_order order); uintptr_t __ovld atomic_fetch_add_explicit(volatile atomic_uintptr_t *object, ptrdiff_t operand, memory_order order, memory_scope scope); -uintptr_t __ovld atomic_fetch_sub(volatile atomic_uintptr_t *object, ptrdiff_t operand); -uintptr_t __ovld atomic_fetch_sub_explicit(volatile atomic_uintptr_t *object, ptrdiff_t operand, memory_order order); uintptr_t __ovld atomic_fetch_sub_explicit(volatile atomic_uintptr_t *object, ptrdiff_t operand, memory_order order, memory_scope scope); - -uintptr_t __ovld atomic_fetch_or(volatile atomic_uintptr_t *object, intptr_t operand); -uintptr_t __ovld atomic_fetch_or_explicit(volatile atomic_uintptr_t *object, intptr_t operand, memory_order order); -uintptr_t __ovld atomic_fetch_or_explicit(volatile atomic_uintptr_t *object, intptr_t operand, memory_order order, memory_scope scope); -uintptr_t __ovld atomic_fetch_xor(volatile atomic_uintptr_t *object, intptr_t operand); -uintptr_t __ovld atomic_fetch_xor_explicit(volatile atomic_uintptr_t *object, intptr_t operand, memory_order order); -uintptr_t __ovld atomic_fetch_xor_explicit(volatile atomic_uintptr_t *object, intptr_t operand, memory_order order, memory_scope scope); -uintptr_t __ovld atomic_fetch_and(volatile atomic_uintptr_t *object, intptr_t operand); -uintptr_t __ovld atomic_fetch_and_explicit(volatile atomic_uintptr_t *object, intptr_t operand, memory_order order); -uintptr_t __ovld atomic_fetch_and_explicit(volatile atomic_uintptr_t *object, intptr_t operand, memory_order order, memory_scope scope); -uintptr_t __ovld atomic_fetch_min(volatile atomic_uintptr_t *object, intptr_t opermax); -uintptr_t __ovld atomic_fetch_min_explicit(volatile atomic_uintptr_t *object, intptr_t opermax, memory_order minder); -uintptr_t __ovld atomic_fetch_min_explicit(volatile atomic_uintptr_t *object, intptr_t opermax, memory_order minder, memory_scope scope); -uintptr_t __ovld atomic_fetch_max(volatile atomic_uintptr_t *object, intptr_t opermax); -uintptr_t __ovld atomic_fetch_max_explicit(volatile atomic_uintptr_t *object, intptr_t opermax, memory_order minder); -uintptr_t __ovld atomic_fetch_max_explicit(volatile atomic_uintptr_t *object, intptr_t opermax, memory_order minder, memory_scope scope); - -intptr_t __ovld atomic_fetch_or(volatile atomic_intptr_t *object, uintptr_t operand); -intptr_t __ovld atomic_fetch_or_explicit(volatile atomic_intptr_t *object, uintptr_t operand, memory_order order); -intptr_t __ovld atomic_fetch_or_explicit(volatile atomic_intptr_t *object, uintptr_t operand, memory_order order, memory_scope scope); -intptr_t __ovld atomic_fetch_xor(volatile atomic_intptr_t *object, uintptr_t operand); -intptr_t __ovld atomic_fetch_xor_explicit(volatile atomic_intptr_t *object, uintptr_t operand, memory_order order); -intptr_t __ovld atomic_fetch_xor_explicit(volatile atomic_intptr_t *object, uintptr_t operand, memory_order order, memory_scope scope); -intptr_t __ovld atomic_fetch_and(volatile atomic_intptr_t *object, uintptr_t operand); -intptr_t __ovld atomic_fetch_and_explicit(volatile atomic_intptr_t *object, uintptr_t operand, memory_order order); -intptr_t __ovld atomic_fetch_and_explicit(volatile atomic_intptr_t *object, uintptr_t operand, memory_order order, memory_scope scope); -intptr_t __ovld atomic_fetch_min(volatile atomic_intptr_t *object, uintptr_t opermax); -intptr_t __ovld atomic_fetch_min_explicit(volatile atomic_intptr_t *object, uintptr_t opermax, memory_order minder); -intptr_t __ovld atomic_fetch_min_explicit(volatile atomic_intptr_t *object, uintptr_t opermax, memory_order minder, memory_scope scope); -intptr_t __ovld atomic_fetch_max(volatile atomic_intptr_t *object, uintptr_t opermax); -intptr_t __ovld atomic_fetch_max_explicit(volatile atomic_intptr_t *object, uintptr_t opermax, memory_order minder); -intptr_t __ovld atomic_fetch_max_explicit(volatile atomic_intptr_t *object, uintptr_t opermax, memory_order minder, memory_scope scope); #endif // atomic_store() +#if defined(__opencl_c_atomic_order_seq_cst) && defined(__opencl_c_atomic_scope_device) void __ovld atomic_store(volatile atomic_int *object, int desired); -void __ovld atomic_store_explicit(volatile atomic_int *object, int desired, memory_order order); -void __ovld atomic_store_explicit(volatile atomic_int *object, int desired, memory_order order, memory_scope scope); void __ovld atomic_store(volatile atomic_uint *object, uint desired); -void __ovld atomic_store_explicit(volatile atomic_uint *object, uint desired, memory_order order); -void __ovld atomic_store_explicit(volatile atomic_uint *object, uint desired, memory_order order, memory_scope scope); void __ovld atomic_store(volatile atomic_float *object, float desired); -void __ovld atomic_store_explicit(volatile atomic_float *object, float desired, memory_order order); -void __ovld atomic_store_explicit(volatile atomic_float *object, float desired, memory_order order, memory_scope scope); + #if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) #ifdef cl_khr_fp64 void __ovld atomic_store(volatile atomic_double *object, double desired); -void __ovld atomic_store_explicit(volatile atomic_double *object, double desired, memory_order order); -void __ovld atomic_store_explicit(volatile atomic_double *object, double desired, memory_order order, memory_scope scope); #endif //cl_khr_fp64 void __ovld atomic_store(volatile atomic_long *object, long desired); -void __ovld atomic_store_explicit(volatile atomic_long *object, long desired, memory_order order); -void __ovld atomic_store_explicit(volatile atomic_long *object, long desired, memory_order order, memory_scope scope); void __ovld atomic_store(volatile atomic_ulong *object, ulong desired); +#endif +#endif + +#if defined(__opencl_c_atomic_scope_device) +void __ovld atomic_store_explicit(volatile atomic_int *object, int desired, memory_order order); +void __ovld atomic_store_explicit(volatile atomic_uint *object, uint desired, memory_order order); +void __ovld atomic_store_explicit(volatile atomic_float *object, float desired, memory_order order); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +void __ovld atomic_store_explicit(volatile atomic_double *object, double desired, memory_order order); +#endif //cl_khr_fp64 +void __ovld atomic_store_explicit(volatile atomic_long *object, long desired, memory_order order); void __ovld atomic_store_explicit(volatile atomic_ulong *object, ulong desired, memory_order order); +#endif +#endif + +void __ovld atomic_store_explicit(volatile atomic_int *object, int desired, memory_order order, memory_scope scope); +void __ovld atomic_store_explicit(volatile atomic_uint *object, uint desired, memory_order order, memory_scope scope); +void __ovld atomic_store_explicit(volatile atomic_float *object, float desired, memory_order order, memory_scope scope); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +void __ovld atomic_store_explicit(volatile atomic_double *object, double desired, memory_order order, memory_scope scope); +#endif //cl_khr_fp64 +void __ovld atomic_store_explicit(volatile atomic_long *object, long desired, memory_order order, memory_scope scope); void __ovld atomic_store_explicit(volatile atomic_ulong *object, ulong desired, memory_order order, memory_scope scope); #endif // atomic_load() - +#if defined(__opencl_c_atomic_order_seq_cst) && defined(__opencl_c_atomic_scope_device) int __ovld atomic_load(volatile atomic_int *object); -int __ovld atomic_load_explicit(volatile atomic_int *object, memory_order order); -int __ovld atomic_load_explicit(volatile atomic_int *object, memory_order order, memory_scope scope); uint __ovld atomic_load(volatile atomic_uint *object); -uint __ovld atomic_load_explicit(volatile atomic_uint *object, memory_order order); -uint __ovld atomic_load_explicit(volatile atomic_uint *object, memory_order order, memory_scope scope); float __ovld atomic_load(volatile atomic_float *object); -float __ovld atomic_load_explicit(volatile atomic_float *object, memory_order order); -float __ovld atomic_load_explicit(volatile atomic_float *object, memory_order order, memory_scope scope); #if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) #ifdef cl_khr_fp64 double __ovld atomic_load(volatile atomic_double *object); -double __ovld atomic_load_explicit(volatile atomic_double *object, memory_order order); -double __ovld atomic_load_explicit(volatile atomic_double *object, memory_order order, memory_scope scope); #endif //cl_khr_fp64 long __ovld atomic_load(volatile atomic_long *object); -long __ovld atomic_load_explicit(volatile atomic_long *object, memory_order order); -long __ovld atomic_load_explicit(volatile atomic_long *object, memory_order order, memory_scope scope); ulong __ovld atomic_load(volatile atomic_ulong *object); +#endif +#endif + +#if defined(__opencl_c_atomic_scope_device) +int __ovld atomic_load_explicit(volatile atomic_int *object, memory_order order); +uint __ovld atomic_load_explicit(volatile atomic_uint *object, memory_order order); +float __ovld atomic_load_explicit(volatile atomic_float *object, memory_order order); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +double __ovld atomic_load_explicit(volatile atomic_double *object, memory_order order); +#endif //cl_khr_fp64 +long __ovld atomic_load_explicit(volatile atomic_long *object, memory_order order); ulong __ovld atomic_load_explicit(volatile atomic_ulong *object, memory_order order); +#endif +#endif + +int __ovld atomic_load_explicit(volatile atomic_int *object, memory_order order, memory_scope scope); +uint __ovld atomic_load_explicit(volatile atomic_uint *object, memory_order order, memory_scope scope); +float __ovld atomic_load_explicit(volatile atomic_float *object, memory_order order, memory_scope scope); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +double __ovld atomic_load_explicit(volatile atomic_double *object, memory_order order, memory_scope scope); +#endif //cl_khr_fp64 +long __ovld atomic_load_explicit(volatile atomic_long *object, memory_order order, memory_scope scope); ulong __ovld atomic_load_explicit(volatile atomic_ulong *object, memory_order order, memory_scope scope); #endif // atomic_exchange() +#if defined(__opencl_c_atomic_order_seq_cst) && defined(__opencl_c_atomic_scope_device) int __ovld atomic_exchange(volatile atomic_int *object, int desired); -int __ovld atomic_exchange_explicit(volatile atomic_int *object, int desired, memory_order order); -int __ovld atomic_exchange_explicit(volatile atomic_int *object, int desired, memory_order order, memory_scope scope); uint __ovld atomic_exchange(volatile atomic_uint *object, uint desired); -uint __ovld atomic_exchange_explicit(volatile atomic_uint *object, uint desired, memory_order order); -uint __ovld atomic_exchange_explicit(volatile atomic_uint *object, uint desired, memory_order order, memory_scope scope); float __ovld atomic_exchange(volatile atomic_float *object, float desired); -float __ovld atomic_exchange_explicit(volatile atomic_float *object, float desired, memory_order order); -float __ovld atomic_exchange_explicit(volatile atomic_float *object, float desired, memory_order order, memory_scope scope); #if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) #ifdef cl_khr_fp64 double __ovld atomic_exchange(volatile atomic_double *object, double desired); -double __ovld atomic_exchange_explicit(volatile atomic_double *object, double desired, memory_order order); -double __ovld atomic_exchange_explicit(volatile atomic_double *object, double desired, memory_order order, memory_scope scope); #endif //cl_khr_fp64 long __ovld atomic_exchange(volatile atomic_long *object, long desired); -long __ovld atomic_exchange_explicit(volatile atomic_long *object, long desired, memory_order order); -long __ovld atomic_exchange_explicit(volatile atomic_long *object, long desired, memory_order order, memory_scope scope); ulong __ovld atomic_exchange(volatile atomic_ulong *object, ulong desired); +#endif +#endif + +#if defined(__opencl_c_atomic_scope_device) +int __ovld atomic_exchange_explicit(volatile atomic_int *object, int desired, memory_order order); +uint __ovld atomic_exchange_explicit(volatile atomic_uint *object, uint desired, memory_order order); +float __ovld atomic_exchange_explicit(volatile atomic_float *object, float desired, memory_order order); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +double __ovld atomic_exchange_explicit(volatile atomic_double *object, double desired, memory_order order); +#endif //cl_khr_fp64 +long __ovld atomic_exchange_explicit(volatile atomic_long *object, long desired, memory_order order); ulong __ovld atomic_exchange_explicit(volatile atomic_ulong *object, ulong desired, memory_order order); +#endif +#endif + +int __ovld atomic_exchange_explicit(volatile atomic_int *object, int desired, memory_order order, memory_scope scope); +uint __ovld atomic_exchange_explicit(volatile atomic_uint *object, uint desired, memory_order order, memory_scope scope); +float __ovld atomic_exchange_explicit(volatile atomic_float *object, float desired, memory_order order, memory_scope scope); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +double __ovld atomic_exchange_explicit(volatile atomic_double *object, double desired, memory_order order, memory_scope scope); +#endif //cl_khr_fp64 +long __ovld atomic_exchange_explicit(volatile atomic_long *object, long desired, memory_order order, memory_scope scope); ulong __ovld atomic_exchange_explicit(volatile atomic_ulong *object, ulong desired, memory_order order, memory_scope scope); #endif // atomic_compare_exchange_strong() and atomic_compare_exchange_weak() - +#if defined(__opencl_c_atomic_order_seq_cst) && defined(__opencl_c_atomic_scope_device) bool __ovld atomic_compare_exchange_strong(volatile atomic_int *object, int *expected, int desired); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_int *object, int *expected, - int desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_int *object, int *expected, - int desired, memory_order success, memory_order failure, memory_scope scope); bool __ovld atomic_compare_exchange_strong(volatile atomic_uint *object, uint *expected, uint desired); +bool __ovld atomic_compare_exchange_weak(volatile atomic_int *object, int *expected, int desired); +bool __ovld atomic_compare_exchange_weak(volatile atomic_uint *object, uint *expected, uint desired); +bool __ovld atomic_compare_exchange_strong(volatile atomic_float *object, float *expected, float desired); +bool __ovld atomic_compare_exchange_weak(volatile atomic_float *object, float *expected, float desired); + +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +bool __ovld atomic_compare_exchange_strong(volatile atomic_double *object, double *expected, double desired); +bool __ovld atomic_compare_exchange_weak(volatile atomic_double *object, double *expected, double desired); +#endif //cl_khr_fp64 +bool __ovld atomic_compare_exchange_strong(volatile atomic_long *object, long *expected, long desired); +bool __ovld atomic_compare_exchange_weak(volatile atomic_long *object, long *expected, long desired); +bool __ovld atomic_compare_exchange_strong(volatile atomic_ulong *object, ulong *expected, ulong desired); +bool __ovld atomic_compare_exchange_weak(volatile atomic_ulong *object, ulong *expected, ulong desired); +#endif +#endif + +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_int *object, int *expected, + int desired, memory_order success, memory_order failure); bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_uint *object, uint *expected, uint desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_uint *object, uint *expected, - uint desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_weak(volatile atomic_int *object, int *expected, int desired); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_int *object, int *expected, int desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_int *object, int *expected, - int desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_weak(volatile atomic_uint *object, uint *expected, uint desired); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_uint *object, uint *expected, uint desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_uint *object, uint *expected, - uint desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_strong(volatile atomic_float *object, float *expected, float desired); bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_float *object, float *expected, float desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_float *object, float *expected, - float desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_weak(volatile atomic_float *object, float *expected, float desired); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_float *object, float *expected, float desired, memory_order success, memory_order failure); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_double *object, double *expected, + double desired, memory_order success, memory_order failure); +bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_double *object, double *expected, + double desired, memory_order success, memory_order failure); +#endif //cl_khr_fp64 +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_long *object, long *expected, + long desired, memory_order success, memory_order failure); +bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_long *object, long *expected, + long desired, memory_order success, memory_order failure); +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_ulong *object, ulong *expected, + ulong desired, memory_order success, memory_order failure); +bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_ulong *object, ulong *expected, + ulong desired, memory_order success, memory_order failure); +#endif + +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_int *object, int *expected, + int desired, memory_order success, memory_order failure, memory_scope scope); +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_uint *object, uint *expected, + uint desired, memory_order success, memory_order failure, memory_scope scope); +bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_int *object, int *expected, + int desired, memory_order success, memory_order failure, memory_scope scope); +bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_uint *object, uint *expected, + uint desired, memory_order success, memory_order failure, memory_scope scope); +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_float *object, float *expected, + float desired, memory_order success, memory_order failure, memory_scope scope); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_float *object, float *expected, float desired, memory_order success, memory_order failure, memory_scope scope); #if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) #ifdef cl_khr_fp64 -bool __ovld atomic_compare_exchange_strong(volatile atomic_double *object, double *expected, double desired); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_double *object, double *expected, - double desired, memory_order success, memory_order failure); bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_double *object, double *expected, double desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_weak(volatile atomic_double *object, double *expected, double desired); -bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_double *object, double *expected, - double desired, memory_order success, memory_order failure); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_double *object, double *expected, double desired, memory_order success, memory_order failure, memory_scope scope); #endif //cl_khr_fp64 -bool __ovld atomic_compare_exchange_strong(volatile atomic_long *object, long *expected, long desired); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_long *object, long *expected, - long desired, memory_order success, memory_order failure); bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_long *object, long *expected, long desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_weak(volatile atomic_long *object, long *expected, long desired); -bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_long *object, long *expected, - long desired, memory_order success, memory_order failure); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_long *object, long *expected, long desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_strong(volatile atomic_ulong *object, ulong *expected, ulong desired); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_ulong *object, ulong *expected, - ulong desired, memory_order success, memory_order failure); bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_ulong *object, ulong *expected, ulong desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_weak(volatile atomic_ulong *object, ulong *expected, ulong desired); -bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_ulong *object, ulong *expected, - ulong desired, memory_order success, memory_order failure); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_ulong *object, ulong *expected, ulong desired, memory_order success, memory_order failure, memory_scope scope); #endif // atomic_flag_test_and_set() and atomic_flag_clear() - +#if defined(__opencl_c_atomic_order_seq_cst) && defined(__opencl_c_atomic_scope_device) bool __ovld atomic_flag_test_and_set(volatile atomic_flag *object); -bool __ovld atomic_flag_test_and_set_explicit(volatile atomic_flag *object, memory_order order); -bool __ovld atomic_flag_test_and_set_explicit(volatile atomic_flag *object, memory_order order, memory_scope scope); void __ovld atomic_flag_clear(volatile atomic_flag *object); -void __ovld atomic_flag_clear_explicit(volatile atomic_flag *object, memory_order order); -void __ovld atomic_flag_clear_explicit(volatile atomic_flag *object, memory_order order, memory_scope scope); +#endif +#if defined(__opencl_c_atomic_scope_device) +bool __ovld atomic_flag_test_and_set_explicit(volatile atomic_flag *object, memory_order order); +void __ovld atomic_flag_clear_explicit(volatile atomic_flag *object, memory_order order); +#endif + +bool __ovld atomic_flag_test_and_set_explicit(volatile atomic_flag *object, memory_order order, memory_scope scope); +void __ovld atomic_flag_clear_explicit(volatile atomic_flag *object, memory_order order, memory_scope scope); #endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) // OpenCL v1.1 s6.11.12, v1.2 s6.12.12, v2.0 s6.13.12 - Miscellaneous Vector Functions @@ -14176,12 +14116,6 @@ half16 __ovld __cnfn shuffle2(half8 x, half8 y, ushort16 mask); half16 __ovld __cnfn shuffle2(half16 x, half16 y, ushort16 mask); #endif //cl_khr_fp16 -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2) -// OpenCL v1.2 s6.12.13, v2.0 s6.13.13 - printf - -int printf(__constant const char* st, ...) __attribute__((format(printf, 1, 2))); -#endif - // OpenCL v1.1 s6.11.3, v1.2 s6.12.14, v2.0 s6.13.14 - Image Read and Write Functions #ifdef cl_khr_gl_msaa_sharing @@ -16130,6 +16064,230 @@ double __ovld sub_group_clustered_reduce_max( double value, uint clustersize ); #endif // cl_khr_subgroup_clustered_reduce +#if defined(cl_khr_extended_bit_ops) +char __ovld __cnfn bitfield_insert(char, char, uint, uint); +uchar __ovld __cnfn bitfield_insert(uchar, uchar, uint, uint); +short __ovld __cnfn bitfield_insert(short, short, uint, uint); +ushort __ovld __cnfn bitfield_insert(ushort, ushort, uint, uint); +int __ovld __cnfn bitfield_insert(int, int, uint, uint); +uint __ovld __cnfn bitfield_insert(uint, uint, uint, uint); +long __ovld __cnfn bitfield_insert(long, long, uint, uint); +ulong __ovld __cnfn bitfield_insert(ulong, ulong, uint, uint); +char2 __ovld __cnfn bitfield_insert(char2, char2, uint, uint); +uchar2 __ovld __cnfn bitfield_insert(uchar2, uchar2, uint, uint); +short2 __ovld __cnfn bitfield_insert(short2, short2, uint, uint); +ushort2 __ovld __cnfn bitfield_insert(ushort2, ushort2, uint, uint); +int2 __ovld __cnfn bitfield_insert(int2, int2, uint, uint); +uint2 __ovld __cnfn bitfield_insert(uint2, uint2, uint, uint); +long2 __ovld __cnfn bitfield_insert(long2, long2, uint, uint); +ulong2 __ovld __cnfn bitfield_insert(ulong2, ulong2, uint, uint); +char3 __ovld __cnfn bitfield_insert(char3, char3, uint, uint); +uchar3 __ovld __cnfn bitfield_insert(uchar3, uchar3, uint, uint); +short3 __ovld __cnfn bitfield_insert(short3, short3, uint, uint); +ushort3 __ovld __cnfn bitfield_insert(ushort3, ushort3, uint, uint); +int3 __ovld __cnfn bitfield_insert(int3, int3, uint, uint); +uint3 __ovld __cnfn bitfield_insert(uint3, uint3, uint, uint); +long3 __ovld __cnfn bitfield_insert(long3, long3, uint, uint); +ulong3 __ovld __cnfn bitfield_insert(ulong3, ulong3, uint, uint); +char4 __ovld __cnfn bitfield_insert(char4, char4, uint, uint); +uchar4 __ovld __cnfn bitfield_insert(uchar4, uchar4, uint, uint); +short4 __ovld __cnfn bitfield_insert(short4, short4, uint, uint); +ushort4 __ovld __cnfn bitfield_insert(ushort4, ushort4, uint, uint); +int4 __ovld __cnfn bitfield_insert(int4, int4, uint, uint); +uint4 __ovld __cnfn bitfield_insert(uint4, uint4, uint, uint); +long4 __ovld __cnfn bitfield_insert(long4, long4, uint, uint); +ulong4 __ovld __cnfn bitfield_insert(ulong4, ulong4, uint, uint); +char8 __ovld __cnfn bitfield_insert(char8, char8, uint, uint); +uchar8 __ovld __cnfn bitfield_insert(uchar8, uchar8, uint, uint); +short8 __ovld __cnfn bitfield_insert(short8, short8, uint, uint); +ushort8 __ovld __cnfn bitfield_insert(ushort8, ushort8, uint, uint); +int8 __ovld __cnfn bitfield_insert(int8, int8, uint, uint); +uint8 __ovld __cnfn bitfield_insert(uint8, uint8, uint, uint); +long8 __ovld __cnfn bitfield_insert(long8, long8, uint, uint); +ulong8 __ovld __cnfn bitfield_insert(ulong8, ulong8, uint, uint); +char16 __ovld __cnfn bitfield_insert(char16, char16, uint, uint); +uchar16 __ovld __cnfn bitfield_insert(uchar16, uchar16, uint, uint); +short16 __ovld __cnfn bitfield_insert(short16, short16, uint, uint); +ushort16 __ovld __cnfn bitfield_insert(ushort16, ushort16, uint, uint); +int16 __ovld __cnfn bitfield_insert(int16, int16, uint, uint); +uint16 __ovld __cnfn bitfield_insert(uint16, uint16, uint, uint); +long16 __ovld __cnfn bitfield_insert(long16, long16, uint, uint); +ulong16 __ovld __cnfn bitfield_insert(ulong16, ulong16, uint, uint); + +char __ovld __cnfn bitfield_extract_signed(char, uint, uint); +short __ovld __cnfn bitfield_extract_signed(short, uint, uint); +int __ovld __cnfn bitfield_extract_signed(int, uint, uint); +long __ovld __cnfn bitfield_extract_signed(long, uint, uint); +char2 __ovld __cnfn bitfield_extract_signed(char2, uint, uint); +short2 __ovld __cnfn bitfield_extract_signed(short2, uint, uint); +int2 __ovld __cnfn bitfield_extract_signed(int2, uint, uint); +long2 __ovld __cnfn bitfield_extract_signed(long2, uint, uint); +char3 __ovld __cnfn bitfield_extract_signed(char3, uint, uint); +short3 __ovld __cnfn bitfield_extract_signed(short3, uint, uint); +int3 __ovld __cnfn bitfield_extract_signed(int3, uint, uint); +long3 __ovld __cnfn bitfield_extract_signed(long3, uint, uint); +char4 __ovld __cnfn bitfield_extract_signed(char4, uint, uint); +short4 __ovld __cnfn bitfield_extract_signed(short4, uint, uint); +int4 __ovld __cnfn bitfield_extract_signed(int4, uint, uint); +long4 __ovld __cnfn bitfield_extract_signed(long4, uint, uint); +char8 __ovld __cnfn bitfield_extract_signed(char8, uint, uint); +short8 __ovld __cnfn bitfield_extract_signed(short8, uint, uint); +int8 __ovld __cnfn bitfield_extract_signed(int8, uint, uint); +long8 __ovld __cnfn bitfield_extract_signed(long8, uint, uint); +char16 __ovld __cnfn bitfield_extract_signed(char16, uint, uint); +short16 __ovld __cnfn bitfield_extract_signed(short16, uint, uint); +int16 __ovld __cnfn bitfield_extract_signed(int16, uint, uint); +long16 __ovld __cnfn bitfield_extract_signed(long16, uint, uint); + +char __ovld __cnfn bitfield_extract_signed(uchar, uint, uint); +short __ovld __cnfn bitfield_extract_signed(ushort, uint, uint); +int __ovld __cnfn bitfield_extract_signed(uint, uint, uint); +long __ovld __cnfn bitfield_extract_signed(ulong, uint, uint); +char2 __ovld __cnfn bitfield_extract_signed(uchar2, uint, uint); +short2 __ovld __cnfn bitfield_extract_signed(ushort2, uint, uint); +int2 __ovld __cnfn bitfield_extract_signed(uint2, uint, uint); +long2 __ovld __cnfn bitfield_extract_signed(ulong2, uint, uint); +char3 __ovld __cnfn bitfield_extract_signed(uchar3, uint, uint); +short3 __ovld __cnfn bitfield_extract_signed(ushort3, uint, uint); +int3 __ovld __cnfn bitfield_extract_signed(uint3, uint, uint); +long3 __ovld __cnfn bitfield_extract_signed(ulong3, uint, uint); +char4 __ovld __cnfn bitfield_extract_signed(uchar4, uint, uint); +short4 __ovld __cnfn bitfield_extract_signed(ushort4, uint, uint); +int4 __ovld __cnfn bitfield_extract_signed(uint4, uint, uint); +long4 __ovld __cnfn bitfield_extract_signed(ulong4, uint, uint); +char8 __ovld __cnfn bitfield_extract_signed(uchar8, uint, uint); +short8 __ovld __cnfn bitfield_extract_signed(ushort8, uint, uint); +int8 __ovld __cnfn bitfield_extract_signed(uint8, uint, uint); +long8 __ovld __cnfn bitfield_extract_signed(ulong8, uint, uint); +char16 __ovld __cnfn bitfield_extract_signed(uchar16, uint, uint); +short16 __ovld __cnfn bitfield_extract_signed(ushort16, uint, uint); +int16 __ovld __cnfn bitfield_extract_signed(uint16, uint, uint); +long16 __ovld __cnfn bitfield_extract_signed(ulong16, uint, uint); + +uchar __ovld __cnfn bitfield_extract_unsigned(char, uint, uint); +ushort __ovld __cnfn bitfield_extract_unsigned(short, uint, uint); +uint __ovld __cnfn bitfield_extract_unsigned(int, uint, uint); +ulong __ovld __cnfn bitfield_extract_unsigned(long, uint, uint); +uchar2 __ovld __cnfn bitfield_extract_unsigned(char2, uint, uint); +ushort2 __ovld __cnfn bitfield_extract_unsigned(short2, uint, uint); +uint2 __ovld __cnfn bitfield_extract_unsigned(int2, uint, uint); +ulong2 __ovld __cnfn bitfield_extract_unsigned(long2, uint, uint); +uchar3 __ovld __cnfn bitfield_extract_unsigned(char3, uint, uint); +ushort3 __ovld __cnfn bitfield_extract_unsigned(short3, uint, uint); +uint3 __ovld __cnfn bitfield_extract_unsigned(int3, uint, uint); +ulong3 __ovld __cnfn bitfield_extract_unsigned(long3, uint, uint); +uchar4 __ovld __cnfn bitfield_extract_unsigned(char4, uint, uint); +ushort4 __ovld __cnfn bitfield_extract_unsigned(short4, uint, uint); +uint4 __ovld __cnfn bitfield_extract_unsigned(int4, uint, uint); +ulong4 __ovld __cnfn bitfield_extract_unsigned(long4, uint, uint); +uchar8 __ovld __cnfn bitfield_extract_unsigned(char8, uint, uint); +ushort8 __ovld __cnfn bitfield_extract_unsigned(short8, uint, uint); +uint8 __ovld __cnfn bitfield_extract_unsigned(int8, uint, uint); +ulong8 __ovld __cnfn bitfield_extract_unsigned(long8, uint, uint); +uchar16 __ovld __cnfn bitfield_extract_unsigned(char16, uint, uint); +ushort16 __ovld __cnfn bitfield_extract_unsigned(short16, uint, uint); +uint16 __ovld __cnfn bitfield_extract_unsigned(int16, uint, uint); +ulong16 __ovld __cnfn bitfield_extract_unsigned(long16, uint, uint); + +uchar __ovld __cnfn bitfield_extract_unsigned(uchar, uint, uint); +ushort __ovld __cnfn bitfield_extract_unsigned(ushort, uint, uint); +uint __ovld __cnfn bitfield_extract_unsigned(uint, uint, uint); +ulong __ovld __cnfn bitfield_extract_unsigned(ulong, uint, uint); +uchar2 __ovld __cnfn bitfield_extract_unsigned(uchar2, uint, uint); +ushort2 __ovld __cnfn bitfield_extract_unsigned(ushort2, uint, uint); +uint2 __ovld __cnfn bitfield_extract_unsigned(uint2, uint, uint); +ulong2 __ovld __cnfn bitfield_extract_unsigned(ulong2, uint, uint); +uchar3 __ovld __cnfn bitfield_extract_unsigned(uchar3, uint, uint); +ushort3 __ovld __cnfn bitfield_extract_unsigned(ushort3, uint, uint); +uint3 __ovld __cnfn bitfield_extract_unsigned(uint3, uint, uint); +ulong3 __ovld __cnfn bitfield_extract_unsigned(ulong3, uint, uint); +uchar4 __ovld __cnfn bitfield_extract_unsigned(uchar4, uint, uint); +ushort4 __ovld __cnfn bitfield_extract_unsigned(ushort4, uint, uint); +uint4 __ovld __cnfn bitfield_extract_unsigned(uint4, uint, uint); +ulong4 __ovld __cnfn bitfield_extract_unsigned(ulong4, uint, uint); +uchar8 __ovld __cnfn bitfield_extract_unsigned(uchar8, uint, uint); +ushort8 __ovld __cnfn bitfield_extract_unsigned(ushort8, uint, uint); +uint8 __ovld __cnfn bitfield_extract_unsigned(uint8, uint, uint); +ulong8 __ovld __cnfn bitfield_extract_unsigned(ulong8, uint, uint); +uchar16 __ovld __cnfn bitfield_extract_unsigned(uchar16, uint, uint); +ushort16 __ovld __cnfn bitfield_extract_unsigned(ushort16, uint, uint); +uint16 __ovld __cnfn bitfield_extract_unsigned(uint16, uint, uint); +ulong16 __ovld __cnfn bitfield_extract_unsigned(ulong16, uint, uint); + +char __ovld __cnfn bit_reverse(char); +uchar __ovld __cnfn bit_reverse(uchar); +short __ovld __cnfn bit_reverse(short); +ushort __ovld __cnfn bit_reverse(ushort); +int __ovld __cnfn bit_reverse(int); +uint __ovld __cnfn bit_reverse(uint); +long __ovld __cnfn bit_reverse(long); +ulong __ovld __cnfn bit_reverse(ulong); +char2 __ovld __cnfn bit_reverse(char2); +uchar2 __ovld __cnfn bit_reverse(uchar2); +short2 __ovld __cnfn bit_reverse(short2); +ushort2 __ovld __cnfn bit_reverse(ushort2); +int2 __ovld __cnfn bit_reverse(int2); +uint2 __ovld __cnfn bit_reverse(uint2); +long2 __ovld __cnfn bit_reverse(long2); +ulong2 __ovld __cnfn bit_reverse(ulong2); +char3 __ovld __cnfn bit_reverse(char3); +uchar3 __ovld __cnfn bit_reverse(uchar3); +short3 __ovld __cnfn bit_reverse(short3); +ushort3 __ovld __cnfn bit_reverse(ushort3); +int3 __ovld __cnfn bit_reverse(int3); +uint3 __ovld __cnfn bit_reverse(uint3); +long3 __ovld __cnfn bit_reverse(long3); +ulong3 __ovld __cnfn bit_reverse(ulong3); +char4 __ovld __cnfn bit_reverse(char4); +uchar4 __ovld __cnfn bit_reverse(uchar4); +short4 __ovld __cnfn bit_reverse(short4); +ushort4 __ovld __cnfn bit_reverse(ushort4); +int4 __ovld __cnfn bit_reverse(int4); +uint4 __ovld __cnfn bit_reverse(uint4); +long4 __ovld __cnfn bit_reverse(long4); +ulong4 __ovld __cnfn bit_reverse(ulong4); +char8 __ovld __cnfn bit_reverse(char8); +uchar8 __ovld __cnfn bit_reverse(uchar8); +short8 __ovld __cnfn bit_reverse(short8); +ushort8 __ovld __cnfn bit_reverse(ushort8); +int8 __ovld __cnfn bit_reverse(int8); +uint8 __ovld __cnfn bit_reverse(uint8); +long8 __ovld __cnfn bit_reverse(long8); +ulong8 __ovld __cnfn bit_reverse(ulong8); +char16 __ovld __cnfn bit_reverse(char16); +uchar16 __ovld __cnfn bit_reverse(uchar16); +short16 __ovld __cnfn bit_reverse(short16); +ushort16 __ovld __cnfn bit_reverse(ushort16); +int16 __ovld __cnfn bit_reverse(int16); +uint16 __ovld __cnfn bit_reverse(uint16); +long16 __ovld __cnfn bit_reverse(long16); +ulong16 __ovld __cnfn bit_reverse(ulong16); +#endif // cl_khr_extended_bit_ops + +#if defined(__opencl_c_integer_dot_product_input_4x8bit) +uint __ovld __cnfn dot(uchar4, uchar4); +int __ovld __cnfn dot(char4, char4); +int __ovld __cnfn dot(uchar4, char4); +int __ovld __cnfn dot(char4, uchar4); + +uint __ovld __cnfn dot_acc_sat(uchar4, uchar4, uint); +int __ovld __cnfn dot_acc_sat(char4, char4, int); +int __ovld __cnfn dot_acc_sat(uchar4, char4, int); +int __ovld __cnfn dot_acc_sat(char4, uchar4, int); +#endif // __opencl_c_integer_dot_product_input_4x8bit + +#if defined(__opencl_c_integer_dot_product_input_4x8bit_packed) +uint __ovld __cnfn dot_4x8packed_uu_uint(uint, uint); +int __ovld __cnfn dot_4x8packed_ss_int(uint, uint); +int __ovld __cnfn dot_4x8packed_us_int(uint, uint); +int __ovld __cnfn dot_4x8packed_su_int(uint, uint); + +uint __ovld __cnfn dot_acc_sat_4x8packed_uu_uint(uint, uint, uint); +int __ovld __cnfn dot_acc_sat_4x8packed_ss_int(uint, uint, int); +int __ovld __cnfn dot_acc_sat_4x8packed_us_int(uint, uint, int); +int __ovld __cnfn dot_acc_sat_4x8packed_su_int(uint, uint, int); +#endif // __opencl_c_integer_dot_product_input_4x8bit_packed + #if defined(cl_intel_subgroups) // Intel-Specific Sub Group Functions float __ovld __conv intel_sub_group_shuffle( float x, uint c ); @@ -17127,31 +17285,23 @@ uint16 __ovld amd_sadw(uint16 src0, uint16 src1, uint16 src2); #endif // cl_amd_media_ops2 #if defined(cl_arm_integer_dot_product_int8) -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_int8 : begin uint __ovld arm_dot(uchar4 a, uchar4 b); int __ovld arm_dot(char4 a, char4 b); -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_int8 : end #endif // defined(cl_arm_integer_dot_product_int8) #if defined(cl_arm_integer_dot_product_accumulate_int8) -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_accumulate_int8 : begin uint __ovld arm_dot_acc(uchar4 a, uchar4 b, uint c); int __ovld arm_dot_acc(char4 a, char4 b, int c); -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_accumulate_int8 : end #endif // defined(cl_arm_integer_dot_product_accumulate_int8) #if defined(cl_arm_integer_dot_product_accumulate_int16) -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_accumulate_int16 : begin uint __ovld arm_dot_acc(ushort2 a, ushort2 b, uint c); int __ovld arm_dot_acc(short2 a, short2 b, int c); -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_accumulate_int16 : end #endif // defined(cl_arm_integer_dot_product_accumulate_int16) #if defined(cl_arm_integer_dot_product_accumulate_saturate_int8) -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_accumulate_saturate_int8 : begin uint __ovld arm_dot_acc_sat(uchar4 a, uchar4 b, uint c); int __ovld arm_dot_acc_sat(char4 a, char4 b, int c); -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_accumulate_saturate_int8 : end #endif // defined(cl_arm_integer_dot_product_accumulate_saturate_int8) // Disable any extensions we may have enabled previously. diff --git a/lib/include/openmp_wrappers/__clang_openmp_device_functions.h b/lib/include/openmp_wrappers/__clang_openmp_device_functions.h index 406c9748e2..953857badf 100644 --- a/lib/include/openmp_wrappers/__clang_openmp_device_functions.h +++ b/lib/include/openmp_wrappers/__clang_openmp_device_functions.h @@ -39,4 +39,46 @@ extern "C" { #pragma omp end declare variant +// Ensure we make `_ZdlPv`, aka. `operator delete(void*)` available without the +// need to `include ` in C++ mode. +#ifdef __cplusplus + +// We require malloc/free. +#include + +#pragma push_macro("OPENMP_NOEXCEPT") +#if __cplusplus >= 201103L +#define OPENMP_NOEXCEPT noexcept +#else +#define OPENMP_NOEXCEPT +#endif + +// Device overrides for non-placement new and delete. +inline void *operator new(__SIZE_TYPE__ size) { + if (size == 0) + size = 1; + return ::malloc(size); +} + +inline void *operator new[](__SIZE_TYPE__ size) { return ::operator new(size); } + +inline void operator delete(void *ptr)OPENMP_NOEXCEPT { ::free(ptr); } + +inline void operator delete[](void *ptr) OPENMP_NOEXCEPT { + ::operator delete(ptr); +} + +// Sized delete, C++14 only. +#if __cplusplus >= 201402L +inline void operator delete(void *ptr, __SIZE_TYPE__ size)OPENMP_NOEXCEPT { + ::operator delete(ptr); +} +inline void operator delete[](void *ptr, __SIZE_TYPE__ size) OPENMP_NOEXCEPT { + ::operator delete(ptr); +} +#endif + +#pragma pop_macro("OPENMP_NOEXCEPT") +#endif + #endif diff --git a/lib/include/openmp_wrappers/complex b/lib/include/openmp_wrappers/complex index 142e526b81..dfd6193c97 100644 --- a/lib/include/openmp_wrappers/complex +++ b/lib/include/openmp_wrappers/complex @@ -17,7 +17,6 @@ // We require std::math functions in the complex builtins below. #include -#define __CUDA__ #define __OPENMP_NVPTX__ #include <__clang_cuda_complex_builtins.h> #undef __OPENMP_NVPTX__ @@ -26,9 +25,6 @@ // Grab the host header too. #include_next - -#ifdef __cplusplus - // If we are compiling against libc++, the macro _LIBCPP_STD_VER should be set // after including above. Since the complex header we use is a // simplified version of the libc++, we don't need it in this case. If we @@ -48,5 +44,3 @@ #pragma omp end declare variant #endif - -#endif diff --git a/lib/include/openmp_wrappers/complex.h b/lib/include/openmp_wrappers/complex.h index 00d278548f..15dc415b81 100644 --- a/lib/include/openmp_wrappers/complex.h +++ b/lib/include/openmp_wrappers/complex.h @@ -17,7 +17,6 @@ // We require math functions in the complex builtins below. #include -#define __CUDA__ #define __OPENMP_NVPTX__ #include <__clang_cuda_complex_builtins.h> #undef __OPENMP_NVPTX__ diff --git a/lib/include/openmp_wrappers/new b/lib/include/openmp_wrappers/new index 1387d925b1..985ddc567f 100644 --- a/lib/include/openmp_wrappers/new +++ b/lib/include/openmp_wrappers/new @@ -9,6 +9,8 @@ #ifndef __CLANG_OPENMP_WRAPPERS_NEW #define __CLANG_OPENMP_WRAPPERS_NEW +// We need the system for the std::nothrow_t. The new/delete operators +// which do not use nothrow_t are provided without the header. #include_next #if defined(__NVPTX__) && defined(_OPENMP) @@ -22,48 +24,24 @@ #define OPENMP_NOEXCEPT #endif -// Device overrides for non-placement new and delete. -inline void *operator new(__SIZE_TYPE__ size) { - if (size == 0) - size = 1; - return ::malloc(size); -} inline void *operator new(__SIZE_TYPE__ size, const std::nothrow_t &) OPENMP_NOEXCEPT { return ::operator new(size); } -inline void *operator new[](__SIZE_TYPE__ size) { return ::operator new(size); } inline void *operator new[](__SIZE_TYPE__ size, const std::nothrow_t &) { return ::operator new(size); } -inline void operator delete(void *ptr)OPENMP_NOEXCEPT { - if (ptr) - ::free(ptr); -} inline void operator delete(void *ptr, const std::nothrow_t &)OPENMP_NOEXCEPT { ::operator delete(ptr); } -inline void operator delete[](void *ptr) OPENMP_NOEXCEPT { - ::operator delete(ptr); -} inline void operator delete[](void *ptr, const std::nothrow_t &) OPENMP_NOEXCEPT { ::operator delete(ptr); } -// Sized delete, C++14 only. -#if __cplusplus >= 201402L -inline void operator delete(void *ptr, __SIZE_TYPE__ size)OPENMP_NOEXCEPT { - ::operator delete(ptr); -} -inline void operator delete[](void *ptr, __SIZE_TYPE__ size) OPENMP_NOEXCEPT { - ::operator delete(ptr); -} -#endif - #pragma pop_macro("OPENMP_NOEXCEPT") #endif diff --git a/lib/include/ppc_wrappers/xmmintrin.h b/lib/include/ppc_wrappers/xmmintrin.h index 0f429fa040..0e45b96769 100644 --- a/lib/include/ppc_wrappers/xmmintrin.h +++ b/lib/include/ppc_wrappers/xmmintrin.h @@ -28,7 +28,7 @@ Most SSE scalar float intrinsic operations can be performed more efficiently as C language float scalar operations or optimized to use vector SIMD operations. We recommend this for new applications. */ -#error "Please read comment above. Use -DNO_WARN_X86_INTRINSICS to disable this error." +#error "Please read comment above. Use -DNO_WARN_X86_INTRINSICS to disable this error." #endif #ifndef _XMMINTRIN_H_INCLUDED @@ -62,14 +62,13 @@ /* The Intel API is flexible enough that we must allow aliasing with other vector types, and their scalar components. */ -typedef float __m128 __attribute__ ((__vector_size__ (16), __may_alias__)); +typedef vector float __m128 __attribute__((__may_alias__)); /* Unaligned version of the same type. */ -typedef float __m128_u __attribute__ ((__vector_size__ (16), __may_alias__, - __aligned__ (1))); +typedef vector float __m128_u __attribute__((__may_alias__, __aligned__(1))); /* Internal data types for implementing the intrinsics. */ -typedef float __v4sf __attribute__ ((__vector_size__ (16))); +typedef vector float __v4sf; /* Create an undefined vector. */ extern __inline __m128 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) diff --git a/lib/include/riscv_vector.h b/lib/include/riscv_vector.h new file mode 100644 index 0000000000..7fc360f9df --- /dev/null +++ b/lib/include/riscv_vector.h @@ -0,0 +1,123865 @@ +/*===---- riscv_vector.h - RISC-V V-extension RVVIntrinsics -------------------=== + * + * + * Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. + * See https://llvm.org/LICENSE.txt for license information. + * SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception + * + *===-----------------------------------------------------------------------=== + */ + +#ifndef __RISCV_VECTOR_H +#define __RISCV_VECTOR_H + +#include +#include + +#ifndef __riscv_vector +#error "Vector intrinsics require the vector extension." +#endif + +#ifdef __cplusplus +extern "C" { +#endif + + +#define vsetvl_e8mf8(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 5) +#define vsetvl_e8mf4(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 6) +#define vsetvl_e8mf2(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 7) +#define vsetvl_e8m1(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 0) +#define vsetvl_e8m2(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 1) +#define vsetvl_e8m4(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 2) +#define vsetvl_e8m8(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 3) + +#define vsetvl_e16mf4(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 6) +#define vsetvl_e16mf2(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 7) +#define vsetvl_e16m1(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 0) +#define vsetvl_e16m2(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 1) +#define vsetvl_e16m4(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 2) +#define vsetvl_e16m8(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 3) + +#define vsetvl_e32mf2(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 7) +#define vsetvl_e32m1(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 0) +#define vsetvl_e32m2(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 1) +#define vsetvl_e32m4(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 2) +#define vsetvl_e32m8(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 3) + +#define vsetvl_e64m1(avl) __builtin_rvv_vsetvli((size_t)(avl), 3, 0) +#define vsetvl_e64m2(avl) __builtin_rvv_vsetvli((size_t)(avl), 3, 1) +#define vsetvl_e64m4(avl) __builtin_rvv_vsetvli((size_t)(avl), 3, 2) +#define vsetvl_e64m8(avl) __builtin_rvv_vsetvli((size_t)(avl), 3, 3) + + +#define vsetvlmax_e8mf8() __builtin_rvv_vsetvlimax(0, 5) +#define vsetvlmax_e8mf4() __builtin_rvv_vsetvlimax(0, 6) +#define vsetvlmax_e8mf2() __builtin_rvv_vsetvlimax(0, 7) +#define vsetvlmax_e8m1() __builtin_rvv_vsetvlimax(0, 0) +#define vsetvlmax_e8m2() __builtin_rvv_vsetvlimax(0, 1) +#define vsetvlmax_e8m4() __builtin_rvv_vsetvlimax(0, 2) +#define vsetvlmax_e8m8() __builtin_rvv_vsetvlimax(0, 3) + +#define vsetvlmax_e16mf4() __builtin_rvv_vsetvlimax(1, 6) +#define vsetvlmax_e16mf2() __builtin_rvv_vsetvlimax(1, 7) +#define vsetvlmax_e16m1() __builtin_rvv_vsetvlimax(1, 0) +#define vsetvlmax_e16m2() __builtin_rvv_vsetvlimax(1, 1) +#define vsetvlmax_e16m4() __builtin_rvv_vsetvlimax(1, 2) +#define vsetvlmax_e16m8() __builtin_rvv_vsetvlimax(1, 3) + +#define vsetvlmax_e32mf2() __builtin_rvv_vsetvlimax(2, 7) +#define vsetvlmax_e32m1() __builtin_rvv_vsetvlimax(2, 0) +#define vsetvlmax_e32m2() __builtin_rvv_vsetvlimax(2, 1) +#define vsetvlmax_e32m4() __builtin_rvv_vsetvlimax(2, 2) +#define vsetvlmax_e32m8() __builtin_rvv_vsetvlimax(2, 3) + +#define vsetvlmax_e64m1() __builtin_rvv_vsetvlimax(3, 0) +#define vsetvlmax_e64m2() __builtin_rvv_vsetvlimax(3, 1) +#define vsetvlmax_e64m4() __builtin_rvv_vsetvlimax(3, 2) +#define vsetvlmax_e64m8() __builtin_rvv_vsetvlimax(3, 3) + +typedef __rvv_bool64_t vbool64_t; +typedef __rvv_bool32_t vbool32_t; +typedef __rvv_bool16_t vbool16_t; +typedef __rvv_bool8_t vbool8_t; +typedef __rvv_bool4_t vbool4_t; +typedef __rvv_bool2_t vbool2_t; +typedef __rvv_bool1_t vbool1_t; +typedef __rvv_int8mf8_t vint8mf8_t; +typedef __rvv_uint8mf8_t vuint8mf8_t; +typedef __rvv_int8mf4_t vint8mf4_t; +typedef __rvv_uint8mf4_t vuint8mf4_t; +typedef __rvv_int8mf2_t vint8mf2_t; +typedef __rvv_uint8mf2_t vuint8mf2_t; +typedef __rvv_int8m1_t vint8m1_t; +typedef __rvv_uint8m1_t vuint8m1_t; +typedef __rvv_int8m2_t vint8m2_t; +typedef __rvv_uint8m2_t vuint8m2_t; +typedef __rvv_int8m4_t vint8m4_t; +typedef __rvv_uint8m4_t vuint8m4_t; +typedef __rvv_int8m8_t vint8m8_t; +typedef __rvv_uint8m8_t vuint8m8_t; +typedef __rvv_int16mf4_t vint16mf4_t; +typedef __rvv_uint16mf4_t vuint16mf4_t; +typedef __rvv_int16mf2_t vint16mf2_t; +typedef __rvv_uint16mf2_t vuint16mf2_t; +typedef __rvv_int16m1_t vint16m1_t; +typedef __rvv_uint16m1_t vuint16m1_t; +typedef __rvv_int16m2_t vint16m2_t; +typedef __rvv_uint16m2_t vuint16m2_t; +typedef __rvv_int16m4_t vint16m4_t; +typedef __rvv_uint16m4_t vuint16m4_t; +typedef __rvv_int16m8_t vint16m8_t; +typedef __rvv_uint16m8_t vuint16m8_t; +typedef __rvv_int32mf2_t vint32mf2_t; +typedef __rvv_uint32mf2_t vuint32mf2_t; +typedef __rvv_int32m1_t vint32m1_t; +typedef __rvv_uint32m1_t vuint32m1_t; +typedef __rvv_int32m2_t vint32m2_t; +typedef __rvv_uint32m2_t vuint32m2_t; +typedef __rvv_int32m4_t vint32m4_t; +typedef __rvv_uint32m4_t vuint32m4_t; +typedef __rvv_int32m8_t vint32m8_t; +typedef __rvv_uint32m8_t vuint32m8_t; +typedef __rvv_int64m1_t vint64m1_t; +typedef __rvv_uint64m1_t vuint64m1_t; +typedef __rvv_int64m2_t vint64m2_t; +typedef __rvv_uint64m2_t vuint64m2_t; +typedef __rvv_int64m4_t vint64m4_t; +typedef __rvv_uint64m4_t vuint64m4_t; +typedef __rvv_int64m8_t vint64m8_t; +typedef __rvv_uint64m8_t vuint64m8_t; +#if defined(__riscv_zfh) +typedef __rvv_float16mf4_t vfloat16mf4_t; +typedef __rvv_float16mf2_t vfloat16mf2_t; +typedef __rvv_float16m1_t vfloat16m1_t; +typedef __rvv_float16m2_t vfloat16m2_t; +typedef __rvv_float16m4_t vfloat16m4_t; +typedef __rvv_float16m8_t vfloat16m8_t; +#endif +#if defined(__riscv_f) +typedef __rvv_float32mf2_t vfloat32mf2_t; +typedef __rvv_float32m1_t vfloat32m1_t; +typedef __rvv_float32m2_t vfloat32m2_t; +typedef __rvv_float32m4_t vfloat32m4_t; +typedef __rvv_float32m8_t vfloat32m8_t; +#endif +#if defined(__riscv_d) +typedef __rvv_float64m1_t vfloat64m1_t; +typedef __rvv_float64m2_t vfloat64m2_t; +typedef __rvv_float64m4_t vfloat64m4_t; +typedef __rvv_float64m8_t vfloat64m8_t; +#endif + +#define vadd_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vadd_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vadd_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vadd_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vadd_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vadd_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vadd_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vadd_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vadd_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vadd_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vadd_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vadd_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vadd_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vadd_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vadd_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vadd_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vadd_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vadd_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vadd_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vadd_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vadd_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vadd_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vadd_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vadd_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vadd_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vadd_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vadd_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vadd_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vadd_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vadd_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vadd_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vadd_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vadd_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vadd_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vadd_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vadd_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vadd_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vadd_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vadd_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vadd_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vadd_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vadd_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vadd_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vadd_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u16mf4((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u16mf2((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u16m1((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u16m2((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u16m4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u16m8((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u32mf2((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u32m1((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u32m2((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u32m4((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u32m8((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u64m1((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u64m2((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u64m4((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u64m8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8m1((const uint8_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8m2((const uint8_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8m4((const uint8_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8m8((const uint8_t *)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (const uint8_t *)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8mf2((const uint8_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8mf4((const uint8_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8mf8((const uint8_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vlse64_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vlse64_v_u64m1((const uint64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vlse64_v_u64m2((const uint64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vlse64_v_u64m4((const uint64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vlse64_v_u64m8((const uint64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vsse16_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsse16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsse16_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_i16m2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsse16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsse16_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_i16m4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsse16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsse16_v_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_i16m8((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vsse16_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_i16m8_m((vbool2_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsse16_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsse16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsse16_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsse16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsse16_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsse16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsse16_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_u16m2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsse16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsse16_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_u16m4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsse16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsse16_v_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_u16m8((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vsse16_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_u16m8_m((vbool2_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsse16_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsse16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsse16_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsse16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsse32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsse32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsse32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_i32m2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsse32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsse32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_i32m4((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsse32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsse32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_i32m8((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsse32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsse32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsse32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsse32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsse32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsse32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_u32m2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsse32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsse32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_u32m4((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsse32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsse32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_u32m8((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsse32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsse32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsse32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsse64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsse64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsse64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_i64m2((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsse64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsse64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_i64m4((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsse64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsse64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_i64m8((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsse64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsse64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsse64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsse64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_u64m2((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsse64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsse64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_u64m4((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsse64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsse64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_u64m8((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsse64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei16_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i8m1((const int8_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i8m2((const int8_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i8m4((const int8_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vluxei16_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxei16_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i8mf2((const int8_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i8mf4((const int8_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i8mf8((const int8_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u8m1((const uint8_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u8m2((const uint8_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u8m4((const uint8_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vluxei16_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxei16_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u8mf2((const uint8_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u8mf4((const uint8_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u8mf8((const uint8_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei32_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i8m1((const int8_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i8m2((const int8_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i8mf2((const int8_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i8mf4((const int8_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i8mf8((const int8_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u8m1((const uint8_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u8m2((const uint8_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u8mf2((const uint8_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u8mf4((const uint8_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u8mf8((const uint8_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i8m1((const int8_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i8mf2((const int8_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i8mf4((const int8_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i8mf8((const int8_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei64_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u8m1((const uint8_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u8mf2((const uint8_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u8mf4((const uint8_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u8mf8((const uint8_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i16m1((const int16_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i16m2((const int16_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i16m4((const int16_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i16m8((const int16_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vluxei8_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxei8_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i16mf2((const int16_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i16mf4((const int16_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei8_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u16m1((const uint16_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u16m2((const uint16_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u16m4((const uint16_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u16m8((const uint16_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vluxei8_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u16mf2((const uint16_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u16mf4((const uint16_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei16_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i16m1((const int16_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i16m2((const int16_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i16m4((const int16_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i16m8((const int16_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vluxei16_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxei16_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i16mf2((const int16_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i16mf4((const int16_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u16m1((const uint16_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u16m2((const uint16_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u16m4((const uint16_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u16m8((const uint16_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vluxei16_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxei16_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u16mf2((const uint16_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u16mf4((const uint16_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei32_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i16m1((const int16_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i16m2((const int16_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i16m4((const int16_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i16mf2((const int16_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i16mf4((const int16_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u16m1((const uint16_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u16m2((const uint16_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u16m4((const uint16_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u16mf2((const uint16_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u16mf4((const uint16_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i16m1((const int16_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i16m2((const int16_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i16mf2((const int16_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i16mf4((const int16_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei64_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u16m1((const uint16_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u16m2((const uint16_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u16mf2((const uint16_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u16mf4((const uint16_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i32m1((const int32_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i32m2((const int32_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i32m4((const int32_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i32m8((const int32_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i32mf2((const int32_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei8_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u32m1((const uint32_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u32m2((const uint32_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u32m4((const uint32_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u32m8((const uint32_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u32mf2((const uint32_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei16_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i32m1((const int32_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i32m2((const int32_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i32m4((const int32_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i32m8((const int32_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i32mf2((const int32_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u32m1((const uint32_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u32m2((const uint32_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u32m4((const uint32_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u32m8((const uint32_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u32mf2((const uint32_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei32_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i32m1((const int32_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i32m2((const int32_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i32m4((const int32_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i32m8((const int32_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i32mf2((const int32_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u32m1((const uint32_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u32m2((const uint32_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u32m4((const uint32_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u32m8((const uint32_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u32mf2((const uint32_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i32m1((const int32_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i32m2((const int32_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i32m4((const int32_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i32mf2((const int32_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei64_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u32m1((const uint32_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u32m2((const uint32_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u32m4((const uint32_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u32mf2((const uint32_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i64m1((const int64_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei8_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i64m2((const int64_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i64m4((const int64_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i64m8((const int64_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u64m1((const uint64_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei8_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u64m2((const uint64_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u64m4((const uint64_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u64m8((const uint64_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i64m1((const int64_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei16_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i64m2((const int64_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i64m4((const int64_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i64m8((const int64_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u64m1((const uint64_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u64m2((const uint64_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u64m4((const uint64_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u64m8((const uint64_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i64m1((const int64_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei32_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i64m2((const int64_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i64m4((const int64_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i64m8((const int64_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u64m1((const uint64_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u64m2((const uint64_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u64m4((const uint64_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u64m8((const uint64_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i64m1((const int64_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei64_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i64m2((const int64_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i64m4((const int64_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i64m8((const int64_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u64m1((const uint64_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei64_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u64m2((const uint64_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u64m4((const uint64_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u64m8((const uint64_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8m1((const int8_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8m2((const int8_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8m4((const int8_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8m8((const int8_t *)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (const int8_t *)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8mf2((const int8_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8mf4((const int8_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8mf8((const int8_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8m1((const uint8_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8m2((const uint8_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8m4((const uint8_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8m8((const uint8_t *)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (const uint8_t *)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8mf2((const uint8_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8mf4((const uint8_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8mf8((const uint8_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei16_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i8m1((const int8_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i8m2((const int8_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i8m4((const int8_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vloxei16_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxei16_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i8mf2((const int8_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i8mf4((const int8_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i8mf8((const int8_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u8m1((const uint8_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u8m2((const uint8_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u8m4((const uint8_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vloxei16_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxei16_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u8mf2((const uint8_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u8mf4((const uint8_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u8mf8((const uint8_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei32_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i8m1((const int8_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i8m2((const int8_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i8mf2((const int8_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i8mf4((const int8_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i8mf8((const int8_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u8m1((const uint8_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u8m2((const uint8_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u8mf2((const uint8_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u8mf4((const uint8_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u8mf8((const uint8_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i8m1((const int8_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i8mf2((const int8_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i8mf4((const int8_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i8mf8((const int8_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei64_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u8m1((const uint8_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u8mf2((const uint8_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u8mf4((const uint8_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u8mf8((const uint8_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i16m1((const int16_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i16m2((const int16_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i16m4((const int16_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i16m8((const int16_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vloxei8_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxei8_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i16mf2((const int16_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i16mf4((const int16_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei8_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u16m1((const uint16_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u16m2((const uint16_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u16m4((const uint16_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u16m8((const uint16_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vloxei8_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxei8_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u16mf2((const uint16_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u16mf4((const uint16_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei16_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i16m1((const int16_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i16m2((const int16_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i16m4((const int16_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i16m8((const int16_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vloxei16_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxei16_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i16mf2((const int16_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i16mf4((const int16_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u16m1((const uint16_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u16m2((const uint16_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u16m4((const uint16_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u16m8((const uint16_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vloxei16_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxei16_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u16mf2((const uint16_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u16mf4((const uint16_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei32_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i16m1((const int16_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i16m2((const int16_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i16m4((const int16_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i16mf2((const int16_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i16mf4((const int16_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u16m1((const uint16_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u16m2((const uint16_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u16m4((const uint16_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u16mf2((const uint16_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u16mf4((const uint16_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i16m1((const int16_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i16m2((const int16_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i16mf2((const int16_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i16mf4((const int16_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei64_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u16m1((const uint16_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u16m2((const uint16_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u16mf2((const uint16_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u16mf4((const uint16_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i32m1((const int32_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i32m2((const int32_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i32m4((const int32_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i32m8((const int32_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i32mf2((const int32_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei8_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u32m1((const uint32_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u32m2((const uint32_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u32m4((const uint32_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u32m8((const uint32_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u32mf2((const uint32_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei16_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i32m1((const int32_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i32m2((const int32_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i32m4((const int32_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i32m8((const int32_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i32mf2((const int32_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u32m1((const uint32_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u32m2((const uint32_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u32m4((const uint32_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u32m8((const uint32_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u32mf2((const uint32_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei32_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i32m1((const int32_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i32m2((const int32_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i32m4((const int32_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i32m8((const int32_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i32mf2((const int32_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u32m1((const uint32_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u32m2((const uint32_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u32m4((const uint32_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u32m8((const uint32_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u32mf2((const uint32_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i32m1((const int32_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i32m2((const int32_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i32m4((const int32_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i32mf2((const int32_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei64_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u32m1((const uint32_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u32m2((const uint32_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u32m4((const uint32_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u32mf2((const uint32_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i64m1((const int64_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei8_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i64m2((const int64_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i64m4((const int64_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i64m8((const int64_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vse8_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vse8_v_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vse8_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8m2((int8_t *)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vse8_v_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vse8_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8m4((int8_t *)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vse8_v_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vse8_v_i8m8(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8m8((int8_t *)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vse8_v_i8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8m8_m((vbool1_t)(op0), (int8_t *)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vse8_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vse8_v_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vse8_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vse8_v_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vse8_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vse8_v_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vloxei8_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u64m1((const uint64_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei8_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u64m2((const uint64_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u64m4((const uint64_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u64m8((const uint64_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i64m1((const int64_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei16_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i64m2((const int64_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i64m4((const int64_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i64m8((const int64_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u64m1((const uint64_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u64m2((const uint64_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u64m4((const uint64_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u64m8((const uint64_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i64m1((const int64_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei32_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i64m2((const int64_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i64m4((const int64_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i64m8((const int64_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u64m1((const uint64_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u64m2((const uint64_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u64m4((const uint64_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u64m8((const uint64_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i64m1((const int64_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei64_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i64m2((const int64_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i64m4((const int64_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i64m8((const int64_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u64m1((const uint64_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei64_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u64m2((const uint64_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u64m4((const uint64_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u64m8((const uint64_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vse8_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vse8_v_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vse8_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vse8_v_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vse8_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8m4((uint8_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vse8_v_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vse8_v_u8m8(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8m8((uint8_t *)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vse8_v_u8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8m8_m((vbool1_t)(op0), (uint8_t *)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vse8_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vse8_v_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vse8_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vse8_v_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vse8_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vse8_v_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vwaddu_wv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u16mf4((vuint16mf4_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u16mf2((vuint16mf2_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u16m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u16m1((vuint16m1_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u16m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u16m2((vuint16m2_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u16m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u16m4((vuint16m4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u16m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u16m8((vuint16m8_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u32mf2((vuint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u32m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u32m1((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u32m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u32m2((vuint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u32m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u32m4((vuint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u32m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u32m8((vuint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u64m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u64m1((vuint64m1_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u64m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u64m2((vuint64m2_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u64m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u64m4((vuint64m4_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u64m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u64m8((vuint64m8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsse8_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsse8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsse8_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8m2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsse8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsse8_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8m4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsse8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsse8_v_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8m8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vsse8_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8m8_m((vbool1_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vsse8_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsse8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsse8_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsse8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsse8_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsse8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i8m4((int8_t *)(op0), (vuint16m8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint16m8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u8m4((uint8_t *)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vadd_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vadd_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vadd_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vadd_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vadd_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vadd_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vadd_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vadd_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vadd_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vadd_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vadd_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vadd_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vadd_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vadd_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vadd_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vadd_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vadd_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vadd_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vadd_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vadd_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vadd_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vadd_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vadd_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vadd_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vadd_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vadd_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vadd_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vadd_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vadd_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vadd_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vadd_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vadd_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vadd_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vadd_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vadd_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vadd_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vadd_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vadd_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vadd_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vadd_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vadd_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vadd_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vadd_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vadd_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vadd_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vadd_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vadd_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vadd_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vadd_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vadd_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vadd_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vadd_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vadd_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vadd_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vadd_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vadd_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vadd_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vadd_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vadd_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vadd_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vadd_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vadd_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vadd_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vadd_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vadd_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vadd_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vadd_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vadd_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vadd_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vadd_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vadd_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vadd_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vadd_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vadd_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vadd_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vadd_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vadd_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vadd_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vadd_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vadd_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vadd_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vadd_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vadd_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vadd_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vadd_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vadd_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vadd_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vadd_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vadd_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vadd_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vadd_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vadd_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vadd_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vadd_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vadd_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vadd_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vadd_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vadd_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vadd_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vadd_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vadd_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vadd_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsub_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vsub_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsub_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vsub_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsub_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vsub_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsub_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vsub_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vsub_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vsub_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsub_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vsub_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsub_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vsub_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsub_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vsub_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsub_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vsub_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsub_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vsub_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsub_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vsub_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsub_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vsub_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsub_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vsub_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsub_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vsub_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsub_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vsub_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsub_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vsub_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsub_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vsub_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsub_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vsub_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsub_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vsub_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsub_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vsub_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsub_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vsub_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsub_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vsub_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsub_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsub_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsub_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsub_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsub_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsub_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsub_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsub_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsub_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsub_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsub_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsub_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsub_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsub_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsub_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsub_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsub_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsub_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsub_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsub_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsub_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsub_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsub_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsub_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsub_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsub_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsub_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsub_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsub_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsub_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vsub_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsub_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vsub_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsub_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vsub_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsub_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vsub_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsub_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vsub_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsub_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vsub_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsub_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vsub_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsub_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vsub_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsub_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vsub_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsub_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vsub_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsub_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vsub_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsub_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vsub_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsub_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vsub_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsub_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vsub_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsub_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vsub_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsub_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vsub_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsub_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vsub_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsub_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vsub_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsub_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vsub_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsub_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vsub_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsub_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vsub_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsub_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vsub_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsub_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsub_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsub_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsub_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsub_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsub_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsub_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsub_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsub_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsub_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsub_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsub_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsub_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsub_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsub_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsub_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsub_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsub_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsub_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsub_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsub_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsub_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsub_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsub_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsub_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsub_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsub_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsub_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsub_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrsub_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrsub_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrsub_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrsub_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrsub_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrsub_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrsub_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrsub_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrsub_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrsub_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrsub_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrsub_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrsub_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrsub_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrsub_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrsub_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrsub_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrsub_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrsub_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrsub_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrsub_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrsub_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrsub_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrsub_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrsub_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrsub_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrsub_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrsub_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrsub_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vrsub_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vrsub_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vrsub_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vrsub_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vrsub_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vrsub_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vrsub_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vrsub_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vrsub_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vrsub_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vrsub_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vrsub_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vrsub_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vrsub_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vrsub_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vrsub_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vrsub_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vrsub_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vrsub_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vrsub_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vrsub_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vrsub_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vrsub_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vrsub_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vrsub_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vrsub_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vrsub_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vrsub_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vrsub_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u16mf4((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u16mf2((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u16m1((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u16m2((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u16m4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u16m8((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u32mf2((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u32m1((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u32m2((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u32m4((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u32m8((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u64m1((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u64m2((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u64m4((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u64m8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u16mf4((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u16mf2((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u16m1((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u16m2((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u16m4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u16m8((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u32mf2((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u32m1((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u32m2((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u32m4((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u32m8((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u64m1((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u64m2((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u64m4((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u64m8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u16mf4((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u16mf2((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u16m1((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u16m2((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u16m4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u16m8((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u32mf2((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u32m1((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u32m2((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u32m4((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u32m8((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u64m1((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u64m2((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u64m4((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u64m8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwadd_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i16mf4((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vwadd_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwadd_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i16mf2((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vwadd_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwadd_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i16m1((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vwadd_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwadd_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i16m2((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vwadd_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwadd_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i16m4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vwadd_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwadd_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i16m8((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vwadd_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwadd_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i32mf2((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vwadd_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwadd_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i32m1((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vwadd_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwadd_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i32m2((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vwadd_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwadd_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i32m4((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vwadd_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwadd_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i32m8((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vwadd_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwadd_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i64m1((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vwadd_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwadd_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i64m2((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vwadd_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwadd_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i64m4((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vwadd_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwadd_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i64m8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vwadd_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vwadd_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i16mf4((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i16mf2((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i16m1((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i16m2((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i16m4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i16m8((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i32mf2((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i32m1((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i32m2((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i32m4((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i32m8((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i64m1((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwadd_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i64m2((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwadd_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i64m4((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwadd_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i64m8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwsub_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i16mf4((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vwsub_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwsub_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i16mf2((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vwsub_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwsub_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i16m1((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vwsub_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwsub_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i16m2((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vwsub_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwsub_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i16m4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vwsub_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwsub_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i16m8((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vwsub_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwsub_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i32mf2((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vwsub_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwsub_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i32m1((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vwsub_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwsub_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i32m2((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vwsub_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwsub_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i32m4((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vwsub_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwsub_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i32m8((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vwsub_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwsub_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i64m1((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vwsub_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwsub_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i64m2((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vwsub_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwsub_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i64m4((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vwsub_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwsub_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i64m8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vwsub_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vwsub_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i16mf4((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i16mf2((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i16m1((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i16m2((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i16m4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i16m8((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i32mf2((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i32m1((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i32m2((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i32m4((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i32m8((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i64m1((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i64m2((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i64m4((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i64m8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u16mf4((vuint16mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u16mf2((vuint16mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u16m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u16m1((vuint16m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u16m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u16m2((vuint16m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u16m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u16m4((vuint16m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u16m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u16m8((vuint16m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u32mf2((vuint32mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u32m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u32m1((vuint32m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u32m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u32m2((vuint32m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u32m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u32m4((vuint32m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u32m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u32m8((vuint32m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u64m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u64m1((vuint64m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u64m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u64m2((vuint64m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u64m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u64m4((vuint64m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u64m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u64m8((vuint64m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u16mf4((vuint16mf4_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u16mf2((vuint16mf2_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u16m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u16m1((vuint16m1_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u16m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u16m2((vuint16m2_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u16m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u16m4((vuint16m4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u16m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u16m8((vuint16m8_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u32mf2((vuint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u32m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u32m1((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u32m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u32m2((vuint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u32m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u32m4((vuint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u32m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u32m8((vuint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u64m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u64m1((vuint64m1_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u64m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u64m2((vuint64m2_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u64m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u64m4((vuint64m4_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u64m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u64m8((vuint64m8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u16mf4((vuint16mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u16mf2((vuint16mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u16m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u16m1((vuint16m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u16m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u16m2((vuint16m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u16m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u16m4((vuint16m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u16m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u16m8((vuint16m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u32mf2((vuint32mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u32m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u32m1((vuint32m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u32m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u32m2((vuint32m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u32m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u32m4((vuint32m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u32m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u32m8((vuint32m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u64m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u64m1((vuint64m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u64m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u64m2((vuint64m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u64m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u64m4((vuint64m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u64m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u64m8((vuint64m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwadd_wv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i16mf4((vint16mf4_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vwadd_wv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwadd_wv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i16mf2((vint16mf2_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vwadd_wv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwadd_wv_i16m1(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i16m1((vint16m1_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vwadd_wv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwadd_wv_i16m2(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i16m2((vint16m2_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vwadd_wv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwadd_wv_i16m4(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i16m4((vint16m4_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vwadd_wv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwadd_wv_i16m8(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i16m8((vint16m8_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vwadd_wv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwadd_wv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i32mf2((vint32mf2_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vwadd_wv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwadd_wv_i32m1(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i32m1((vint32m1_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vwadd_wv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwadd_wv_i32m2(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i32m2((vint32m2_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vwadd_wv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwadd_wv_i32m4(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i32m4((vint32m4_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vwadd_wv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwadd_wv_i32m8(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i32m8((vint32m8_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vwadd_wv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwadd_wv_i64m1(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i64m1((vint64m1_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vwadd_wv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwadd_wv_i64m2(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i64m2((vint64m2_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vwadd_wv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwadd_wv_i64m4(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i64m4((vint64m4_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vwadd_wv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwadd_wv_i64m8(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i64m8((vint64m8_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vwadd_wv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vwadd_wx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i16mf4((vint16mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_wx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_wx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i16mf2((vint16mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_wx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_wx_i16m1(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i16m1((vint16m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_wx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_wx_i16m2(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i16m2((vint16m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_wx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_wx_i16m4(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i16m4((vint16m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_wx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_wx_i16m8(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i16m8((vint16m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_wx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_wx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i32mf2((vint32mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_wx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_wx_i32m1(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i32m1((vint32m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_wx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_wx_i32m2(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i32m2((vint32m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_wx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_wx_i32m4(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i32m4((vint32m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_wx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_wx_i32m8(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i32m8((vint32m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_wx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_wx_i64m1(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i64m1((vint64m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_wx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwadd_wx_i64m2(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i64m2((vint64m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_wx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwadd_wx_i64m4(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i64m4((vint64m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_wx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwadd_wx_i64m8(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i64m8((vint64m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_wx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_wv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i16mf4((vint16mf4_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vwsub_wv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwsub_wv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i16mf2((vint16mf2_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vwsub_wv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwsub_wv_i16m1(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i16m1((vint16m1_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vwsub_wv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwsub_wv_i16m2(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i16m2((vint16m2_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vwsub_wv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwsub_wv_i16m4(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i16m4((vint16m4_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vwsub_wv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwsub_wv_i16m8(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i16m8((vint16m8_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vwsub_wv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwsub_wv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i32mf2((vint32mf2_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vwsub_wv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwsub_wv_i32m1(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i32m1((vint32m1_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vwsub_wv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwsub_wv_i32m2(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i32m2((vint32m2_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vwsub_wv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwsub_wv_i32m4(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i32m4((vint32m4_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vwsub_wv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwsub_wv_i32m8(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i32m8((vint32m8_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vwsub_wv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwsub_wv_i64m1(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i64m1((vint64m1_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vwsub_wv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwsub_wv_i64m2(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i64m2((vint64m2_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vwsub_wv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwsub_wv_i64m4(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i64m4((vint64m4_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vwsub_wv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwsub_wv_i64m8(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i64m8((vint64m8_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vwsub_wv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vwsub_wx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i16mf4((vint16mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_wx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_wx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i16mf2((vint16mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_wx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_wx_i16m1(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i16m1((vint16m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_wx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_wx_i16m2(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i16m2((vint16m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_wx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_wx_i16m4(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i16m4((vint16m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_wx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_wx_i16m8(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i16m8((vint16m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_wx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_wx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i32mf2((vint32mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_wx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_wx_i32m1(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i32m1((vint32m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_wx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_wx_i32m2(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i32m2((vint32m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_wx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_wx_i32m4(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i32m4((vint32m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_wx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_wx_i32m8(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i32m8((vint32m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_wx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_wx_i64m1(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i64m1((vint64m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_wx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_wx_i64m2(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i64m2((vint64m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_wx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_wx_i64m4(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i64m4((vint64m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_wx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_wx_i64m8(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i64m8((vint64m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_wx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwcvt_x_x_v_i16m2(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i16m2((vint8m1_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i16m4(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i16m4((vint8m2_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i16m8(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i16m8((vint8m4_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i16m1(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i16m1((vint8mf2_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i16mf2(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i16mf2((vint8mf4_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i16mf4(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i16mf4((vint8mf8_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i32m2(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i32m2((vint16m1_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i32m4(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i32m4((vint16m2_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i32m8(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i32m8((vint16m4_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i32m1(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i32m1((vint16mf2_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i32mf2(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i32mf2((vint16mf4_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i64m2(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i64m2((vint32m1_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i64m4(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i64m4((vint32m2_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i64m8(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i64m8((vint32m4_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i64m1(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i64m1((vint32mf2_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vadc_vvm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vvm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vvm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vadc_vvm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vvm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vvm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vvm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8m1((vint8m1_t)(op0), (int8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8m2((vint8m2_t)(op0), (int8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8m4((vint8m4_t)(op0), (int8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8m8((vint8m8_t)(op0), (int8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i16m1((vint16m1_t)(op0), (int16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i16m2((vint16m2_t)(op0), (int16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i16m4((vint16m4_t)(op0), (int16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vxm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i16m8((vint16m8_t)(op0), (int16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vxm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i32m1((vint32m1_t)(op0), (int32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i32m2((vint32m2_t)(op0), (int32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i32m4((vint32m4_t)(op0), (int32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i32m8((vint32m8_t)(op0), (int32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vxm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i64m1((vint64m1_t)(op0), (int64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i64m2((vint64m2_t)(op0), (int64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i64m4((vint64m4_t)(op0), (int64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i64m8((vint64m8_t)(op0), (int64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vvm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vvm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vvm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vxm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vxm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vxm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmadc_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmadc_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmadc_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmadc_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmadc_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmadc_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmadc_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmadc_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmadc_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmadc_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmadc_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmadc_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmadc_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmadc_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmadc_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmadc_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmadc_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmadc_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmadc_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmadc_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsuxei8_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i16m4((int16_t *)(op0), (vuint8m2_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint8m2_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i16m8((int16_t *)(op0), (vuint8m4_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i16m8_m((vbool2_t)(op0), (int16_t *)(op1), (vuint8m4_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmadc_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmadc_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmadc_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmadc_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmadc_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmadc_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmadc_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmadc_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmadc_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmadc_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmadc_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmadc_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmadc_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmadc_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmadc_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmadc_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmadc_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmadc_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmadc_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmadc_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmadc_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsbc_vvm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8m1((vint8m1_t)(op0), (int8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8m2((vint8m2_t)(op0), (int8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8m4((vint8m4_t)(op0), (int8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8m8((vint8m8_t)(op0), (int8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i16m1((vint16m1_t)(op0), (int16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i16m2((vint16m2_t)(op0), (int16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i16m4((vint16m4_t)(op0), (int16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i16m8((vint16m8_t)(op0), (int16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i32m1((vint32m1_t)(op0), (int32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i32m2((vint32m2_t)(op0), (int32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i32m4((vint32m4_t)(op0), (int32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i32m8((vint32m8_t)(op0), (int32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i64m1((vint64m1_t)(op0), (int64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i64m2((vint64m2_t)(op0), (int64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i64m4((vint64m4_t)(op0), (int64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i64m8((vint64m8_t)(op0), (int64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u16m4((uint16_t *)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint8m2_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u16m8((uint16_t *)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u16m8_m((vbool2_t)(op0), (uint16_t *)(op1), (vuint8m4_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmsbc_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vand_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vand_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vand_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vand_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vand_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vand_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vand_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vand_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vand_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vand_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vand_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vand_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vand_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vand_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vand_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vand_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vand_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vand_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vand_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vand_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vand_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vand_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vand_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vand_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vand_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vand_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vand_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vand_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vand_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vand_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vand_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vand_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vand_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vand_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vand_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vand_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vand_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vand_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vand_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vand_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vand_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vand_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vand_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vand_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vand_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vand_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vand_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vand_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vand_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vand_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vand_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vand_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vand_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vand_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vand_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vand_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vand_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vand_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vand_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vand_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vand_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vand_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vand_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vand_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vand_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vand_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vand_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vand_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vand_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vand_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vand_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vand_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vand_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vand_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vand_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vand_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vand_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vand_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vand_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vand_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vand_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vand_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vand_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vand_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vand_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vand_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vand_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vand_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vand_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vand_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vand_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vand_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vand_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vand_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vand_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vand_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vand_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vand_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vand_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vand_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vand_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vand_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vand_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vand_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vand_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vand_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vand_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vand_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vand_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vand_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vand_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vand_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vand_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vand_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vand_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vand_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vand_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vand_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vand_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vand_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vand_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vand_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vand_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vand_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vand_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vand_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vand_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vand_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vand_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vand_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vand_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vand_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vand_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vand_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vand_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vand_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vand_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vand_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vand_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vand_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vand_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vand_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vand_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vand_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vand_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vand_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vand_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vand_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vand_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vand_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vand_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vand_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vxor_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vxor_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vxor_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vxor_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vxor_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vxor_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vxor_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vxor_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vxor_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vxor_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vxor_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vxor_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vxor_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vxor_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vxor_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vxor_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vxor_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vxor_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vxor_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vxor_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vxor_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vxor_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vxor_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vxor_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vxor_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vxor_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vxor_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vxor_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vxor_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vxor_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vxor_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vxor_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vxor_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vxor_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vxor_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vxor_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vxor_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vxor_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vxor_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vxor_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vxor_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vxor_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vxor_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vxor_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vxor_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vxor_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vxor_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vxor_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vxor_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vxor_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vxor_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vxor_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vxor_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vxor_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vxor_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vxor_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vxor_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vxor_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vxor_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vxor_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vxor_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vxor_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vxor_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vxor_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vxor_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vxor_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vxor_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vxor_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vxor_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vxor_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vxor_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vxor_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vxor_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i16m4((int16_t *)(op0), (vuint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i16m8((int16_t *)(op0), (vuint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i16m8_m((vbool2_t)(op0), (int16_t *)(op1), (vuint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vxor_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vxor_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vxor_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vxor_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vxor_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vxor_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vxor_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vxor_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vxor_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vxor_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vxor_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vxor_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vxor_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vxor_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vxor_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vxor_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vxor_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vxor_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vxor_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vxor_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vxor_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vxor_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vxor_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vxor_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vxor_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vxor_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vxor_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vxor_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vxor_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vxor_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vxor_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vxor_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vxor_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vxor_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vxor_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vxor_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vxor_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vxor_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vxor_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vxor_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vxor_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vxor_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vxor_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vxor_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vxor_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vxor_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vxor_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vxor_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vxor_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vxor_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vxor_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vxor_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vxor_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vxor_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vxor_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vxor_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vxor_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vxor_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vxor_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vxor_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vxor_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vxor_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vxor_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vxor_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vxor_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vxor_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vxor_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vxor_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vxor_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vxor_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vxor_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vxor_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vxor_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vor_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vor_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vor_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vor_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vor_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vor_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vor_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vor_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vor_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vor_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vor_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vor_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vor_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vor_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vor_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vor_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vor_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vor_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vor_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vor_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vor_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vor_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vor_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vor_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vor_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vor_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vor_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vor_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vor_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vor_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vor_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vor_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vor_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vor_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vor_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vor_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vor_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vor_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vor_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vor_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vor_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vor_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vor_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vor_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vor_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vor_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vor_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vor_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vor_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vor_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vor_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vor_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vor_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vor_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vor_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vor_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vor_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vor_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vor_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vor_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vor_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vor_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vor_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vor_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vor_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vor_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vor_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vor_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vor_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vor_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vor_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vor_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vor_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vor_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vor_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vor_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vor_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vor_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vor_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vor_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vor_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vor_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vor_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vor_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vor_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vor_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vor_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vor_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vor_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vor_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vor_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vor_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vor_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vor_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vor_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vor_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vor_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vor_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vor_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vor_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vor_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vor_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vor_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vor_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vor_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vor_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vor_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vor_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vor_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vor_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vor_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vor_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vor_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vor_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vor_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vor_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vor_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vor_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vor_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vor_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vor_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vor_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vor_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vor_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vor_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vor_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vor_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vor_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vor_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vor_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vor_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vor_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vor_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vor_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vor_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vor_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vor_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vor_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vor_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vor_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vor_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vor_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vor_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vor_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vor_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vor_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vor_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vor_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vor_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vor_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vor_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vor_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsll_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8m1((vint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vsll_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsll_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8m2((vint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vsll_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsll_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8m4((vint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vsll_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsll_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8m8((vint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vsll_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsll_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8mf2((vint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vsll_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsll_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8mf4((vint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vsll_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsll_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8mf8((vint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vsll_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsll_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i16m1((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vsll_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsll_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i16m2((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vsll_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsll_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i16m4((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vsll_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsll_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i16m8((vint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vsll_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsll_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i16mf2((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vsll_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsll_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i16mf4((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vsll_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsll_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i32m1((vint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vsll_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsll_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i32m2((vint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vsll_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsll_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i32m4((vint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vsll_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsll_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i32m8((vint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vsll_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsll_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i32mf2((vint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vsll_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsll_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i64m1((vint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vsll_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsll_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i64m2((vint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vsll_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsll_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i64m4((vint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vsll_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsll_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i64m8((vint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vsll_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsll_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8m1((vint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8m2((vint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8m4((vint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8m8((vint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8mf2((vint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8mf4((vint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8mf8((vint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i16m1((vint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i16m2((vint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i16m4((vint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i16m8((vint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i16mf2((vint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i16mf4((vint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i32m1((vint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i32m2((vint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i32m4((vint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i32m8((vint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i32mf2((vint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i64m1((vint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i64m2((vint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i64m4((vint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i64m8((vint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vsll_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsll_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vsll_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsll_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vsll_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsll_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vsll_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsll_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vsll_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsll_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vsll_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsll_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vsll_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsll_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vsll_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsll_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vsll_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsll_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vsll_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsll_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vsll_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsll_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vsll_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsll_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vsll_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsll_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vsll_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsll_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vsll_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsll_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vsll_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsll_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vsll_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsll_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vsll_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsll_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vsll_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsll_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vsll_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsll_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vsll_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsll_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vsll_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsll_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8m1((vuint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8m2((vuint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8m4((vuint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8m8((vuint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8mf2((vuint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8mf4((vuint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8mf8((vuint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u16m1((vuint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u16m2((vuint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u16m4((vuint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u16m8((vuint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u16mf2((vuint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u16mf4((vuint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u32m1((vuint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u32m2((vuint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u32m4((vuint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u32m8((vuint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u32mf2((vuint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u64m1((vuint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u64m2((vuint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u64m4((vuint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u64m8((vuint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u16m4((uint16_t *)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u16m8((uint16_t *)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u16m8_m((vbool2_t)(op0), (uint16_t *)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsrl_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vsrl_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsrl_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vsrl_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vsrl_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsrl_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vsrl_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vsrl_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsrl_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vsrl_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vsrl_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsrl_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vsrl_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsrl_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vsrl_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vsrl_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8m1((vuint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8m2((vuint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8m4((vuint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8m8((vuint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8mf2((vuint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8mf4((vuint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8mf8((vuint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u16m1((vuint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u16m2((vuint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u16m4((vuint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u16m8((vuint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u16mf2((vuint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u16mf4((vuint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u32m1((vuint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u32m2((vuint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u32m4((vuint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u32m8((vuint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u32mf2((vuint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u64m1((vuint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u64m2((vuint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u64m4((vuint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u64m8((vuint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8m1((vint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vsra_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsra_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8m2((vint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vsra_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsra_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8m4((vint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vsra_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsra_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8m8((vint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vsra_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsra_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8mf2((vint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vsra_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsra_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8mf4((vint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vsra_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsra_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8mf8((vint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vsra_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsra_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i16m1((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vsra_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsra_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i16m2((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vsra_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsra_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i16m4((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vsra_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsra_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i16m8((vint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vsra_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsra_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i16mf2((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vsra_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsra_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i16mf4((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vsra_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsra_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i32m1((vint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vsra_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsra_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i32m2((vint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vsra_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsra_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i32m4((vint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vsra_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsra_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i32m8((vint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vsra_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsra_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i32mf2((vint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vsra_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsra_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i64m1((vint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vsra_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsra_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i64m2((vint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vsra_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsra_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i64m4((vint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vsra_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsra_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i64m8((vint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vsra_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsra_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8m1((vint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8m2((vint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8m4((vint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8m8((vint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8mf2((vint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8mf4((vint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8mf8((vint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i16m1((vint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i16m2((vint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i16m4((vint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i16m8((vint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i16mf2((vint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i16mf4((vint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i32m1((vint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i32m2((vint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i32m4((vint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i32m8((vint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i32mf2((vint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i64m1((vint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i64m2((vint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i64m4((vint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i64m8((vint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u8m1(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u8m1((vuint16m2_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u8m2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u8m2((vuint16m4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u8m4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u8m4((vuint16m8_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u8mf2((vuint16m1_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u8mf4((vuint16mf2_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u8mf8((vuint16mf4_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u16m1(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u16m1((vuint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u16m2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u16m2((vuint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u16m4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u16m4((vuint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u16mf2((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u16mf4((vuint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u32m1(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u32m1((vuint64m2_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u32m2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u32m2((vuint64m4_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u32m4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u32m4((vuint64m8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u32mf2((vuint64m1_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u8m1(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u8m1((vuint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u8m2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u8m2((vuint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u8m4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u8m4((vuint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u8mf2((vuint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u8mf4((vuint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u8mf8((vuint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u16m1(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u16m1((vuint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u16m2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u16m2((vuint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u16m4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u16m4((vuint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u16mf2((vuint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u16mf4((vuint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u32m1(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u32m1((vuint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u32m2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u32m2((vuint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u32m4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u32m4((vuint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u32mf2((vuint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wv_i8m1(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i8m1((vint16m2_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vnsra_wv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnsra_wv_i8m2(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i8m2((vint16m4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vnsra_wv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnsra_wv_i8m4(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i8m4((vint16m8_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vnsra_wv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnsra_wv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i8mf2((vint16m1_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vnsra_wv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnsra_wv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i8mf4((vint16mf2_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vnsra_wv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnsra_wv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i8mf8((vint16mf4_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vnsra_wv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnsra_wv_i16m1(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i16m1((vint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vnsra_wv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnsra_wv_i16m2(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i16m2((vint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vnsra_wv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnsra_wv_i16m4(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i16m4((vint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vnsra_wv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnsra_wv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i16mf2((vint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vnsra_wv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnsra_wv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i16mf4((vint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vnsra_wv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnsra_wv_i32m1(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i32m1((vint64m2_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vnsra_wv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnsra_wv_i32m2(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i32m2((vint64m4_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vnsra_wv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnsra_wv_i32m4(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i32m4((vint64m8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vnsra_wv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnsra_wv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i32mf2((vint64m1_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vnsra_wv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnsra_wx_i8m1(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i8m1((vint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i8m2(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i8m2((vint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i8m4(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i8m4((vint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i8mf2((vint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i8mf4((vint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i8mf8((vint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i16m1(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i16m1((vint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i16m2(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i16m2((vint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i16m4(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i16m4((vint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i16mf2((vint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i16mf4((vint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i32m1(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i32m1((vint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i32m2(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i32m2((vint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i32m4(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i32m4((vint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i32mf2((vint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vncvt_x_x_w_u8m1(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u8m1((vuint16m2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u8m2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u8m2((vuint16m4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u8m4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u8m4((vuint16m8_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u8mf2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u8mf2((vuint16m1_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u8mf4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u8mf4((vuint16mf2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u8mf8(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u8mf8((vuint16mf4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u16m1(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u16m1((vuint32m2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u16m2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u16m2((vuint32m4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u16m4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u16m4((vuint32m8_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u16mf2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u16mf2((vuint32m1_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u16mf4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u16mf4((vuint32mf2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u32m1(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u32m1((vuint64m2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u32m2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u32m2((vuint64m4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u32m4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u32m4((vuint64m8_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u32mf2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u32mf2((vuint64m1_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vmseq_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmseq_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmseq_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmseq_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmseq_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmseq_vv_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmseq_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmseq_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsse8_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsse8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsse8_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8m2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsse8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsse8_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8m4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vsse8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsse8_v_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8m8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vsse8_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8m8_m((vbool1_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsse8_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsse8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsse8_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsse8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsse8_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsse8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i16m4((int16_t *)(op0), (vuint32m8_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint32m8_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmseq_vx_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmseq_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmseq_vx_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmseq_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmseq_vx_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmseq_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmseq_vx_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmseq_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmseq_vx_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmseq_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmseq_vx_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmseq_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmseq_vx_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmseq_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmseq_vx_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmseq_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmseq_vx_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmseq_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmseq_vx_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmseq_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmseq_vx_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmseq_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmseq_vx_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmseq_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmseq_vx_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmseq_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmseq_vx_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmseq_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmseq_vx_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmseq_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmseq_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmseq_vv_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmseq_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmseq_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmseq_vx_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmseq_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmseq_vx_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmseq_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmseq_vx_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmseq_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmseq_vx_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmseq_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmseq_vx_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmseq_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmseq_vx_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmseq_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmseq_vx_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmseq_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmseq_vx_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmseq_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmseq_vx_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmseq_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmseq_vx_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmseq_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmseq_vx_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmseq_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmseq_vx_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmseq_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmseq_vx_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmseq_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmseq_vx_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmseq_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmseq_vx_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmsne_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmsne_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmsne_vv_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmsne_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmsne_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsne_vx_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsne_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsne_vx_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsne_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsne_vx_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsne_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsne_vx_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsne_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsne_vx_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsne_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsne_vx_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsne_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsne_vx_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsne_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsne_vx_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsne_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsne_vx_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsne_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsne_vx_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsne_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsne_vx_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsne_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsne_vx_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsne_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsne_vx_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsne_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsne_vx_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsne_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsne_vx_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmsne_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmsne_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmsne_vv_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmsne_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmsne_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsne_vx_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsne_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsne_vx_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsne_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsne_vx_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsne_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsne_vx_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsne_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsne_vx_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsne_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsne_vx_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsne_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsne_vx_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsne_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsne_vx_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsne_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsne_vx_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsne_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsne_vx_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsne_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsne_vx_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsne_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsne_vx_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsne_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsne_vx_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsne_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsne_vx_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsne_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsne_vx_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmslt_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmslt_vv_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmslt_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmslt_vv_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmslt_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmslt_vv_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmslt_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmslt_vv_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmslt_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmslt_vv_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmslt_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmslt_vv_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmslt_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmslt_vv_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmslt_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmslt_vv_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmslt_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmslt_vv_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmslt_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmslt_vv_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u16m4((uint16_t *)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmslt_vx_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmslt_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmslt_vx_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmslt_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmslt_vx_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmslt_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmslt_vx_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmslt_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmslt_vx_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmslt_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmslt_vx_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmslt_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmslt_vx_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmslt_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmslt_vx_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmslt_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmslt_vx_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmslt_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmslt_vx_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmslt_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmslt_vx_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmslt_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmslt_vx_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmslt_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmslt_vx_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmslt_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmslt_vx_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmslt_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmslt_vx_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmsle_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmsle_vv_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmsle_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmsle_vv_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmsle_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmsle_vv_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmsle_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmsle_vv_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmsle_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmsle_vv_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmsle_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmsle_vv_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmsle_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmsle_vv_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmsle_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmsle_vv_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmsle_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmsle_vv_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmsle_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmsle_vv_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsle_vx_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsle_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsle_vx_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsle_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsle_vx_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsle_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsle_vx_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsle_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsle_vx_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsle_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsle_vx_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsle_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsle_vx_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsle_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsle_vx_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsle_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsle_vx_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsle_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsle_vx_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsle_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsle_vx_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsle_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsle_vx_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsle_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsle_vx_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsle_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsle_vx_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsle_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsle_vx_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmsge_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmsge_vv_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmsge_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmsge_vv_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmsge_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmsge_vv_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmsge_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmsge_vv_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmsge_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmsge_vv_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmsge_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmsge_vv_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmsge_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmsge_vv_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmsge_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmsge_vv_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmsge_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmsge_vv_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmsge_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmsge_vv_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsge_vx_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsge_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsge_vx_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsge_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsge_vx_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsge_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsge_vx_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsge_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsge_vx_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsge_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsge_vx_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsge_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsge_vx_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsge_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsge_vx_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsge_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsge_vx_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsge_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsge_vx_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsge_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsge_vx_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsge_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsge_vx_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsge_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsge_vx_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsge_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsge_vx_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsge_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsge_vx_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vminu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vminu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vminu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vminu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vminu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vminu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vminu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vminu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vminu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vminu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vminu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vminu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vminu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vminu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vminu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vminu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vminu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vminu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vminu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vminu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vminu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vminu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vminu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vminu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vminu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vminu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vminu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vminu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vminu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vminu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vminu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vminu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vminu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vminu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vminu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vminu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vminu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vminu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vminu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vminu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vminu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vminu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vminu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vminu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vminu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vminu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vminu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vminu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vminu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vminu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vminu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vminu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vminu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vminu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vminu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vminu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vminu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vminu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vminu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vminu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vminu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vminu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vminu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vminu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vminu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vminu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vminu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vminu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vminu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vminu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vminu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vminu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vminu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmin_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmin_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmin_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmin_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmin_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmin_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmin_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmin_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmin_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmin_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmin_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmin_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmin_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmin_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmin_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmin_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmin_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmin_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmin_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmin_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmin_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmin_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmin_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmin_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmin_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmin_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmin_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmin_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmin_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmin_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmin_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmin_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmin_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmin_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmin_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmin_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmin_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmin_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmin_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmin_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmin_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmin_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmin_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmin_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmin_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmin_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmin_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmin_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmin_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmin_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmin_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmin_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmin_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmin_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmin_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmin_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmin_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmin_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmin_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmin_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmin_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmin_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmin_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmin_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmin_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmin_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmin_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmin_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmin_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmin_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmin_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmin_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmin_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmax_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmax_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmax_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmax_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmax_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmax_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmax_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmax_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmax_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmax_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmax_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmax_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmax_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmax_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmax_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmax_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmax_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmax_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmax_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmax_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmax_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmax_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmax_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmax_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmax_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmax_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmax_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmax_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmax_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmax_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmax_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmax_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmax_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmax_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmax_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmax_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmax_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmax_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmax_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmax_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmax_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmax_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmax_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmax_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmax_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmax_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmax_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmax_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmax_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmax_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmax_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmax_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmax_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmax_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmax_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmax_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmax_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmax_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmax_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmax_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmax_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmax_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmax_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmax_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmax_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmax_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmax_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmax_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmax_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmax_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmax_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmax_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmax_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmax_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmul_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmul_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmul_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmul_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmul_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmul_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmul_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmul_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmul_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmul_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmul_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmul_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmul_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmul_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmul_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmul_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmul_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmul_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmul_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmul_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmul_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmul_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmul_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmul_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmul_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmul_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmul_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmul_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmul_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmul_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmul_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmul_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmul_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmul_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmul_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmul_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmul_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmul_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmul_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmul_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmul_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmul_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmul_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmul_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmul_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmul_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmul_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmul_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmul_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmul_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmul_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmul_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmul_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmul_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmul_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmul_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmul_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmul_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmul_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmul_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmul_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmul_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmul_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmul_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmul_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmul_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmul_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmul_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmul_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmul_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmul_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmul_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmul_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmul_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmul_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmul_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmul_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmul_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmul_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmul_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmul_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmul_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmul_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmul_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmul_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmul_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmul_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmul_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmul_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmul_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmul_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmul_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmul_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmul_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmul_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmul_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmul_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmul_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmul_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmul_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmul_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmul_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmul_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmul_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmul_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmul_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmul_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmul_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmul_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmul_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmul_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmul_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmul_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmul_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmul_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmul_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmul_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmul_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmul_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmul_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmul_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmul_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmul_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmul_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmul_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmul_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmul_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmul_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmul_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmul_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmul_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmul_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmul_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmul_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmul_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmul_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmul_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmul_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmul_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmul_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmul_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmul_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmul_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmul_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmul_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmul_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmulh_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmulh_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmulh_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmulh_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmulh_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmulh_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmulh_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmulh_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmulh_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmulh_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmulh_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmulh_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmulh_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmulh_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmulh_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmulh_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmulh_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmulh_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmulh_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmulh_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmulh_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmulh_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmulh_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmulh_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmulh_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmulh_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmulh_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmulh_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmulh_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmulh_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmulh_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmulh_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmulh_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmulh_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmulh_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmulh_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmulh_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmulh_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmulh_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmulh_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmulh_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmulh_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmulh_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmulh_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmulh_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmulh_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmulh_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmulh_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmulh_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8m1((vint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8m2((vint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8m4((vint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8m8((vint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8mf2((vint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8mf4((vint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8mf8((vint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i16m1((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i16m2((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i16m4((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i16m8((vint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i16mf2((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i16mf4((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i32m1((vint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i32m2((vint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i32m4((vint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i32m8((vint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i32mf2((vint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i64m1((vint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i64m2((vint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i64m4((vint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i64m8((vint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8m1((vint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8m2((vint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8m4((vint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8m8((vint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8mf2((vint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8mf4((vint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8mf8((vint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i16m1((vint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i16m2((vint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i16m4((vint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i16m8((vint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i16mf2((vint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i16mf4((vint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i32m1((vint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i32m2((vint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i32m4((vint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i32m8((vint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i32mf2((vint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i64m1((vint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i64m2((vint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i64m4((vint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i64m8((vint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vdivu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vdivu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vdivu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vdivu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vdivu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vdivu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vdivu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vdivu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vdivu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vdivu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vdivu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vdivu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vdivu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vdivu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vdivu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vdivu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vdivu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vdivu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vdivu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vdivu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vdivu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vdivu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vdivu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vdivu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vdivu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vdivu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vdivu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vdivu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vdivu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vdivu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vdivu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vdivu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vdivu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vdivu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vdivu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vdivu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vdivu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vdivu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vdivu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vdivu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vdivu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vdivu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vdivu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vdivu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vdivu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vdivu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vdivu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vdivu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vdivu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vdiv_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vdiv_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vdiv_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vdiv_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vdiv_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vdiv_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vdiv_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vdiv_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vdiv_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vdiv_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vdiv_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vdiv_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vdiv_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vdiv_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vdiv_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vdiv_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vdiv_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vdiv_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vdiv_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vdiv_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vdiv_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vdiv_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vdiv_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vdiv_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vdiv_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vdiv_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vdiv_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vdiv_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vdiv_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vdiv_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vdiv_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vdiv_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vdiv_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vdiv_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vdiv_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vdiv_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vdiv_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vdiv_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vdiv_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vdiv_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vdiv_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vdiv_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vdiv_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vdiv_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vdiv_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vdiv_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vdiv_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vdiv_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vdiv_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vremu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vremu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vremu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vremu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vremu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vremu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vremu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vremu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vremu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vremu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vremu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vremu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vremu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vremu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vremu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vremu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vremu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vremu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vremu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vremu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vremu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vremu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vremu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vremu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vremu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vremu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vremu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vremu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vremu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vremu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vremu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vremu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vremu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vremu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vremu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vremu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vremu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vremu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vremu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vremu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vremu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vremu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vremu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vremu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vremu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vremu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vremu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vremu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vremu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vremu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vremu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vremu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vremu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vremu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vremu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vremu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vremu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vremu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vremu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vremu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vremu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vremu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vremu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vremu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vremu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vremu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vremu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vremu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vremu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vremu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vremu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vremu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vremu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vrem_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vrem_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vrem_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vrem_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vrem_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vrem_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vrem_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vrem_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vrem_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vrem_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vrem_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vrem_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vrem_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vrem_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vrem_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vrem_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vrem_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vrem_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vrem_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vrem_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vrem_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vrem_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vrem_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vrem_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vrem_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vrem_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vrem_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vrem_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vrem_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vrem_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vrem_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vrem_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vrem_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vrem_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vrem_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vrem_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vrem_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vrem_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vrem_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vrem_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vrem_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vrem_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vrem_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vrem_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrem_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrem_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrem_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrem_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrem_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrem_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrem_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrem_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrem_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrem_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrem_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrem_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrem_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrem_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrem_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrem_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrem_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrem_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrem_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrem_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrem_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrem_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrem_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrem_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrem_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrem_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrem_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrem_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrem_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vwmul_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i16mf4((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vwmul_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwmul_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i16mf2((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vwmul_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwmul_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i16m1((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vwmul_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwmul_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i16m2((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vwmul_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwmul_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i16m4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vwmul_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwmul_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i16m8((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vwmul_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwmul_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i32mf2((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vwmul_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwmul_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i32m1((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vwmul_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwmul_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i32m2((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vwmul_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwmul_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i32m4((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vwmul_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwmul_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i32m8((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vwmul_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwmul_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i64m1((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vwmul_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwmul_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i64m2((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vwmul_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwmul_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i64m4((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vwmul_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwmul_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i64m8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vwmul_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmul_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i16mf4((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwmul_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwmul_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i16mf2((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwmul_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwmul_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i16m1((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwmul_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwmul_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i16m2((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwmul_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwmul_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i16m4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwmul_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwmul_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i16m8((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwmul_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwmul_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i32mf2((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwmul_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwmul_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i32m1((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwmul_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwmul_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i32m2((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwmul_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwmul_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i32m4((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwmul_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwmul_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i32m8((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwmul_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwmul_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i64m1((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwmul_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwmul_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i64m2((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwmul_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwmul_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i64m4((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwmul_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwmul_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i64m8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwmul_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u16mf4((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u16mf2((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u16m1((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u16m2((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u16m4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u16m8((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u32mf2((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u32m1((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u32m2((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u32m4((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u32m8((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u64m1((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u64m2((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u64m4((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u64m8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u16mf4((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u16mf2((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u16m1((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u16m2((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u16m4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u16m8((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u32mf2((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u32m1((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u32m2((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u32m4((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u32m8((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u64m1((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u64m2((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u64m4((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u64m8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i16mf4((vint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i16mf2((vint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i16m1((vint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i16m2((vint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i16m4((vint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i16m8((vint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i32mf2((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i32m1((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i32m2((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i32m4((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i32m8((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i64m1((vint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i64m2((vint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i64m4((vint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i64m8((vint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i16mf4((vint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i16mf2((vint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i16m1((vint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i16m2((vint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i16m4((vint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i16m8((vint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i32mf2((vint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i32m1((vint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i32m2((vint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i32m4((vint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i32m8((vint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i64m1((vint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i64m2((vint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i64m4((vint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i64m8((vint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmacc_vv_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vmacc_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmacc_vv_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (int8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (int8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (int8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (int8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (int8_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (int8_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (int8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmacc_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (int16_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (int16_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (int16_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (int16_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (int16_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vmacc_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (int16_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmacc_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (int32_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (int32_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (int32_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (int32_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (int32_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (int64_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (int64_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (int64_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (int64_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmacc_vv_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vmacc_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmacc_vv_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (uint8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (uint8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (uint8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (uint8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (uint8_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (uint8_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (uint8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmacc_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (uint16_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (uint16_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (uint16_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (uint16_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (uint16_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vmacc_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (uint16_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmacc_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (uint32_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (uint32_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (uint32_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (uint32_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (uint32_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (uint64_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (uint64_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (uint64_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (uint64_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (int8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (int8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (int8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (int8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (int8_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (int8_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (int8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (int16_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (int16_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (int16_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (int16_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (int16_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (int16_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (int32_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (int32_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (int32_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (int32_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (int32_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (int64_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (int64_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (int64_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (int64_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (uint8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (uint8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (uint8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (uint8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (uint8_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (uint8_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (uint8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (uint16_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (uint16_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (uint16_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (uint16_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (uint16_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (uint16_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (uint32_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (uint32_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (uint32_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (uint32_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (uint32_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (uint64_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (uint64_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (uint64_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (uint64_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmadd_vv_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vmadd_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmadd_vv_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (int8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (int8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (int8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (int8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (int8_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (int8_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (int8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmadd_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (int16_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (int16_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (int16_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (int16_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (int16_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vmadd_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (int16_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmadd_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (int32_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (int32_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (int32_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (int32_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (int32_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (int64_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (int64_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (int64_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (int64_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmadd_vv_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vmadd_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmadd_vv_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (uint8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (uint8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (uint8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (uint8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (uint8_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (uint8_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (uint8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmadd_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (uint16_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (uint16_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (uint16_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (uint16_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (uint16_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vmadd_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (uint16_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmadd_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (uint32_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (uint32_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (uint32_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (uint32_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (uint32_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (uint64_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (uint64_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (uint64_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (uint64_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (int8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (int8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (int8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (int8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (int8_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (int8_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (int8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (int16_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (int16_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (int16_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (int16_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (int16_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (int16_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (int32_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (int32_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (int32_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (int32_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (int32_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (int64_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (int64_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (int64_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (int64_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (uint8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (uint8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (uint8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (uint8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (uint8_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (uint8_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (uint8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (uint16_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (uint16_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (uint16_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (uint16_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (uint16_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (uint16_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (uint32_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (uint32_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (uint32_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (uint32_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (uint32_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (uint64_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (uint64_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (uint64_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (uint64_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u16mf4((vuint16mf4_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u16mf2((vuint16mf2_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u16m1((vuint16m1_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u16m2((vuint16m2_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u16m4((vuint16m4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u16m8((vuint16m8_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u32mf2((vuint32mf2_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u32m1((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u32m2((vuint32m2_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u32m4((vuint32m4_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u32m8((vuint32m8_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u64m1((vuint64m1_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u64m2((vuint64m2_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u64m4((vuint64m4_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u64m8((vuint64m8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u16mf4((vuint16mf4_t)(op0), (uint8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (uint8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u16mf2((vuint16mf2_t)(op0), (uint8_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (uint8_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u16m1((vuint16m1_t)(op0), (uint8_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (uint8_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u16m2((vuint16m2_t)(op0), (uint8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (uint8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u16m4((vuint16m4_t)(op0), (uint8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (uint8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u16m8((vuint16m8_t)(op0), (uint8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (uint8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u32mf2((vuint32mf2_t)(op0), (uint16_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (uint16_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u32m1((vuint32m1_t)(op0), (uint16_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (uint16_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u32m2((vuint32m2_t)(op0), (uint16_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (uint16_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u32m4((vuint32m4_t)(op0), (uint16_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (uint16_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u32m8((vuint32m8_t)(op0), (uint16_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (uint16_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u64m1((vuint64m1_t)(op0), (uint32_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (uint32_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u64m2((vuint64m2_t)(op0), (uint32_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (uint32_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u64m4((vuint64m4_t)(op0), (uint32_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (uint32_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u64m8((vuint64m8_t)(op0), (uint32_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (uint32_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i16mf4((vint16mf4_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i16mf2((vint16mf2_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i16m1((vint16m1_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i16m2((vint16m2_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i16m4((vint16m4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i16m8((vint16m8_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i32mf2((vint32mf2_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i32m1((vint32m1_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i32m2((vint32m2_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i32m4((vint32m4_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i32m8((vint32m8_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i64m1((vint64m1_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i64m2((vint64m2_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i64m4((vint64m4_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i64m8((vint64m8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i16mf4((vint16mf4_t)(op0), (int8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (int8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i16mf2((vint16mf2_t)(op0), (int8_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (int8_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i16m1((vint16m1_t)(op0), (int8_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (int8_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i16m2((vint16m2_t)(op0), (int8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (int8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i16m4((vint16m4_t)(op0), (int8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (int8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i16m8((vint16m8_t)(op0), (int8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (int8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i32mf2((vint32mf2_t)(op0), (int16_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (int16_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i32m1((vint32m1_t)(op0), (int16_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (int16_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i32m2((vint32m2_t)(op0), (int16_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (int16_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i32m4((vint32m4_t)(op0), (int16_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (int16_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i32m8((vint32m8_t)(op0), (int16_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (int16_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i64m1((vint64m1_t)(op0), (int32_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (int32_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i64m2((vint64m2_t)(op0), (int32_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (int32_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i64m4((vint64m4_t)(op0), (int32_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (int32_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i64m8((vint64m8_t)(op0), (int32_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (int32_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i16mf4((vint16mf4_t)(op0), (vint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i16mf2((vint16mf2_t)(op0), (vint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i16m1((vint16m1_t)(op0), (vint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i16m2((vint16m2_t)(op0), (vint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i16m4((vint16m4_t)(op0), (vint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i16m8((vint16m8_t)(op0), (vint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i32mf2((vint32mf2_t)(op0), (vint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i32m1((vint32m1_t)(op0), (vint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i32m2((vint32m2_t)(op0), (vint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i32m4((vint32m4_t)(op0), (vint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i32m8((vint32m8_t)(op0), (vint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i64m1((vint64m1_t)(op0), (vint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i64m2((vint64m2_t)(op0), (vint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i64m4((vint64m4_t)(op0), (vint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i64m8((vint64m8_t)(op0), (vint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i16mf4((vint16mf4_t)(op0), (int8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (int8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i16mf2((vint16mf2_t)(op0), (int8_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (int8_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i16m1((vint16m1_t)(op0), (int8_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (int8_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i16m2((vint16m2_t)(op0), (int8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (int8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i16m4((vint16m4_t)(op0), (int8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (int8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i16m8((vint16m8_t)(op0), (int8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (int8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i32mf2((vint32mf2_t)(op0), (int16_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (int16_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i32m1((vint32m1_t)(op0), (int16_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (int16_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i32m2((vint32m2_t)(op0), (int16_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (int16_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i32m4((vint32m4_t)(op0), (int16_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (int16_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i32m8((vint32m8_t)(op0), (int16_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (int16_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i64m1((vint64m1_t)(op0), (int32_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (int32_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i64m2((vint64m2_t)(op0), (int32_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (int32_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i64m4((vint64m4_t)(op0), (int32_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (int32_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i64m8((vint64m8_t)(op0), (int32_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (int32_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i16mf4((vint16mf4_t)(op0), (uint8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (uint8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i16mf2((vint16mf2_t)(op0), (uint8_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (uint8_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i16m1((vint16m1_t)(op0), (uint8_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (uint8_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i16m2((vint16m2_t)(op0), (uint8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (uint8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i16m4((vint16m4_t)(op0), (uint8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (uint8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i16m8((vint16m8_t)(op0), (uint8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (uint8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i32mf2((vint32mf2_t)(op0), (uint16_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (uint16_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i32m1((vint32m1_t)(op0), (uint16_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (uint16_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i32m2((vint32m2_t)(op0), (uint16_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (uint16_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i32m4((vint32m4_t)(op0), (uint16_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (uint16_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i32m8((vint32m8_t)(op0), (uint16_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (uint16_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i64m1((vint64m1_t)(op0), (uint32_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (uint32_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i64m2((vint64m2_t)(op0), (uint32_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (uint32_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i64m4((vint64m4_t)(op0), (uint32_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (uint32_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i64m8((vint64m8_t)(op0), (uint32_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (uint32_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmerge_vvm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8m1((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8m2((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8m4((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8m8((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8mf2((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8mf4((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8mf8((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i16m1((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i16m2((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i16m4((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i16m8((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i16mf2((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i16mf4((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i32m1((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i32m2((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i32m4((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i32m8((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i32mf2((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i64m1((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i64m2((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i64m4((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i64m8((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8m1((vbool8_t)(op0), (vint8m1_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8m2((vbool4_t)(op0), (vint8m2_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8m4((vbool2_t)(op0), (vint8m4_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8m8((vbool1_t)(op0), (vint8m8_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8mf2((vbool16_t)(op0), (vint8mf2_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8mf4((vbool32_t)(op0), (vint8mf4_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8mf8((vbool64_t)(op0), (vint8mf8_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i16m1((vbool16_t)(op0), (vint16m1_t)(op1), (int16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i16m2((vbool8_t)(op0), (vint16m2_t)(op1), (int16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i16m4((vbool4_t)(op0), (vint16m4_t)(op1), (int16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i16m8((vbool2_t)(op0), (vint16m8_t)(op1), (int16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i16mf2((vbool32_t)(op0), (vint16mf2_t)(op1), (int16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i16mf4((vbool64_t)(op0), (vint16mf4_t)(op1), (int16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i32m1((vbool32_t)(op0), (vint32m1_t)(op1), (int32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i32m2((vbool16_t)(op0), (vint32m2_t)(op1), (int32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i32m4((vbool8_t)(op0), (vint32m4_t)(op1), (int32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i32m8((vbool4_t)(op0), (vint32m8_t)(op1), (int32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i32mf2((vbool64_t)(op0), (vint32mf2_t)(op1), (int32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i64m1((vbool64_t)(op0), (vint64m1_t)(op1), (int64_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i64m2((vbool32_t)(op0), (vint64m2_t)(op1), (int64_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i64m4((vbool16_t)(op0), (vint64m4_t)(op1), (int64_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i64m8((vbool8_t)(op0), (vint64m8_t)(op1), (int64_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmerge_vvm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8m1((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8m2((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8m4((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8m8((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8mf2((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8mf4((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8mf8((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u16m1((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u16m2((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u16m4((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u16m8((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u16mf2((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u16mf4((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u32m1((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u32m2((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u32m4((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u32m8((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u32mf2((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u64m1((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u64m2((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u64m4((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u64m8((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8m1((vbool8_t)(op0), (vuint8m1_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8m2((vbool4_t)(op0), (vuint8m2_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8m4((vbool2_t)(op0), (vuint8m4_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8m8((vbool1_t)(op0), (vuint8m8_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8mf2((vbool16_t)(op0), (vuint8mf2_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8mf4((vbool32_t)(op0), (vuint8mf4_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8mf8((vbool64_t)(op0), (vuint8mf8_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u16m1((vbool16_t)(op0), (vuint16m1_t)(op1), (uint16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u16m2((vbool8_t)(op0), (vuint16m2_t)(op1), (uint16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u16m4((vbool4_t)(op0), (vuint16m4_t)(op1), (uint16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u16m8((vbool2_t)(op0), (vuint16m8_t)(op1), (uint16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u16mf2((vbool32_t)(op0), (vuint16mf2_t)(op1), (uint16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u16mf4((vbool64_t)(op0), (vuint16mf4_t)(op1), (uint16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u32m1((vbool32_t)(op0), (vuint32m1_t)(op1), (uint32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u32m2((vbool16_t)(op0), (vuint32m2_t)(op1), (uint32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u32m4((vbool8_t)(op0), (vuint32m4_t)(op1), (uint32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u32m8((vbool4_t)(op0), (vuint32m8_t)(op1), (uint32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u32mf2((vbool64_t)(op0), (vuint32mf2_t)(op1), (uint32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u64m1((vbool64_t)(op0), (vuint64m1_t)(op1), (uint64_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u64m2((vbool32_t)(op0), (vuint64m2_t)(op1), (uint64_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u64m4((vbool16_t)(op0), (vuint64m4_t)(op1), (uint64_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u64m8((vbool8_t)(op0), (vuint64m8_t)(op1), (uint64_t)(op2), (size_t)(op3)) +#define vmv_v_v_u8m1(op0, op1) \ +__builtin_rvv_vmv_v_v_u8m1((vuint8m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_u8m2(op0, op1) \ +__builtin_rvv_vmv_v_v_u8m2((vuint8m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u8m4(op0, op1) \ +__builtin_rvv_vmv_v_v_u8m4((vuint8m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_u8m8(op0, op1) \ +__builtin_rvv_vmv_v_v_u8m8((vuint8m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_u8mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_u8mf2((vuint8mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u8mf4(op0, op1) \ +__builtin_rvv_vmv_v_v_u8mf4((vuint8mf4_t)(op0), (size_t)(op1)) +#define vmv_v_v_u8mf8(op0, op1) \ +__builtin_rvv_vmv_v_v_u8mf8((vuint8mf8_t)(op0), (size_t)(op1)) +#define vmv_v_v_u16m1(op0, op1) \ +__builtin_rvv_vmv_v_v_u16m1((vuint16m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_u16m2(op0, op1) \ +__builtin_rvv_vmv_v_v_u16m2((vuint16m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u16m4(op0, op1) \ +__builtin_rvv_vmv_v_v_u16m4((vuint16m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_u16m8(op0, op1) \ +__builtin_rvv_vmv_v_v_u16m8((vuint16m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_u16mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_u16mf2((vuint16mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u16mf4(op0, op1) \ +__builtin_rvv_vmv_v_v_u16mf4((vuint16mf4_t)(op0), (size_t)(op1)) +#define vmv_v_v_u32m1(op0, op1) \ +__builtin_rvv_vmv_v_v_u32m1((vuint32m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_u32m2(op0, op1) \ +__builtin_rvv_vmv_v_v_u32m2((vuint32m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u32m4(op0, op1) \ +__builtin_rvv_vmv_v_v_u32m4((vuint32m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_u32m8(op0, op1) \ +__builtin_rvv_vmv_v_v_u32m8((vuint32m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_u32mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_u32mf2((vuint32mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u64m1(op0, op1) \ +__builtin_rvv_vmv_v_v_u64m1((vuint64m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_u64m2(op0, op1) \ +__builtin_rvv_vmv_v_v_u64m2((vuint64m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u64m4(op0, op1) \ +__builtin_rvv_vmv_v_v_u64m4((vuint64m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_u64m8(op0, op1) \ +__builtin_rvv_vmv_v_v_u64m8((vuint64m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8m1(op0, op1) \ +__builtin_rvv_vmv_v_v_i8m1((vint8m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8m2(op0, op1) \ +__builtin_rvv_vmv_v_v_i8m2((vint8m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8m4(op0, op1) \ +__builtin_rvv_vmv_v_v_i8m4((vint8m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8m8(op0, op1) \ +__builtin_rvv_vmv_v_v_i8m8((vint8m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_i8mf2((vint8mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8mf4(op0, op1) \ +__builtin_rvv_vmv_v_v_i8mf4((vint8mf4_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8mf8(op0, op1) \ +__builtin_rvv_vmv_v_v_i8mf8((vint8mf8_t)(op0), (size_t)(op1)) +#define vmv_v_v_i16m1(op0, op1) \ +__builtin_rvv_vmv_v_v_i16m1((vint16m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_i16m2(op0, op1) \ +__builtin_rvv_vmv_v_v_i16m2((vint16m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i16m4(op0, op1) \ +__builtin_rvv_vmv_v_v_i16m4((vint16m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_i16m8(op0, op1) \ +__builtin_rvv_vmv_v_v_i16m8((vint16m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_i16mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_i16mf2((vint16mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i16mf4(op0, op1) \ +__builtin_rvv_vmv_v_v_i16mf4((vint16mf4_t)(op0), (size_t)(op1)) +#define vmv_v_v_i32m1(op0, op1) \ +__builtin_rvv_vmv_v_v_i32m1((vint32m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_i32m2(op0, op1) \ +__builtin_rvv_vmv_v_v_i32m2((vint32m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i32m4(op0, op1) \ +__builtin_rvv_vmv_v_v_i32m4((vint32m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_i32m8(op0, op1) \ +__builtin_rvv_vmv_v_v_i32m8((vint32m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_i32mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_i32mf2((vint32mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i64m1(op0, op1) \ +__builtin_rvv_vmv_v_v_i64m1((vint64m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_i64m2(op0, op1) \ +__builtin_rvv_vmv_v_v_i64m2((vint64m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i64m4(op0, op1) \ +__builtin_rvv_vmv_v_v_i64m4((vint64m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_i64m8(op0, op1) \ +__builtin_rvv_vmv_v_v_i64m8((vint64m8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8m1(op0, op1) \ +__builtin_rvv_vmv_v_x_i8m1((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8m2(op0, op1) \ +__builtin_rvv_vmv_v_x_i8m2((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8m4(op0, op1) \ +__builtin_rvv_vmv_v_x_i8m4((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8m8(op0, op1) \ +__builtin_rvv_vmv_v_x_i8m8((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8mf2(op0, op1) \ +__builtin_rvv_vmv_v_x_i8mf2((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8mf4(op0, op1) \ +__builtin_rvv_vmv_v_x_i8mf4((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8mf8(op0, op1) \ +__builtin_rvv_vmv_v_x_i8mf8((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i16m1(op0, op1) \ +__builtin_rvv_vmv_v_x_i16m1((int16_t)(op0), (size_t)(op1)) +#define vmv_v_x_i16m2(op0, op1) \ +__builtin_rvv_vmv_v_x_i16m2((int16_t)(op0), (size_t)(op1)) +#define vmv_v_x_i16m4(op0, op1) \ +__builtin_rvv_vmv_v_x_i16m4((int16_t)(op0), (size_t)(op1)) +#define vmv_v_x_i16m8(op0, op1) \ +__builtin_rvv_vmv_v_x_i16m8((int16_t)(op0), (size_t)(op1)) +#define vmv_v_x_i16mf2(op0, op1) \ +__builtin_rvv_vmv_v_x_i16mf2((int16_t)(op0), (size_t)(op1)) +#define vmv_v_x_i16mf4(op0, op1) \ +__builtin_rvv_vmv_v_x_i16mf4((int16_t)(op0), (size_t)(op1)) +#define vmv_v_x_i32m1(op0, op1) \ +__builtin_rvv_vmv_v_x_i32m1((int32_t)(op0), (size_t)(op1)) +#define vmv_v_x_i32m2(op0, op1) \ +__builtin_rvv_vmv_v_x_i32m2((int32_t)(op0), (size_t)(op1)) +#define vmv_v_x_i32m4(op0, op1) \ +__builtin_rvv_vmv_v_x_i32m4((int32_t)(op0), (size_t)(op1)) +#define vmv_v_x_i32m8(op0, op1) \ +__builtin_rvv_vmv_v_x_i32m8((int32_t)(op0), (size_t)(op1)) +#define vmv_v_x_i32mf2(op0, op1) \ +__builtin_rvv_vmv_v_x_i32mf2((int32_t)(op0), (size_t)(op1)) +#define vmv_v_x_i64m1(op0, op1) \ +__builtin_rvv_vmv_v_x_i64m1((int64_t)(op0), (size_t)(op1)) +#define vmv_v_x_i64m2(op0, op1) \ +__builtin_rvv_vmv_v_x_i64m2((int64_t)(op0), (size_t)(op1)) +#define vmv_v_x_i64m4(op0, op1) \ +__builtin_rvv_vmv_v_x_i64m4((int64_t)(op0), (size_t)(op1)) +#define vmv_v_x_i64m8(op0, op1) \ +__builtin_rvv_vmv_v_x_i64m8((int64_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8m1(op0, op1) \ +__builtin_rvv_vmv_v_x_u8m1((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8m2(op0, op1) \ +__builtin_rvv_vmv_v_x_u8m2((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8m4(op0, op1) \ +__builtin_rvv_vmv_v_x_u8m4((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8m8(op0, op1) \ +__builtin_rvv_vmv_v_x_u8m8((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8mf2(op0, op1) \ +__builtin_rvv_vmv_v_x_u8mf2((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8mf4(op0, op1) \ +__builtin_rvv_vmv_v_x_u8mf4((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8mf8(op0, op1) \ +__builtin_rvv_vmv_v_x_u8mf8((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u16m1(op0, op1) \ +__builtin_rvv_vmv_v_x_u16m1((uint16_t)(op0), (size_t)(op1)) +#define vmv_v_x_u16m2(op0, op1) \ +__builtin_rvv_vmv_v_x_u16m2((uint16_t)(op0), (size_t)(op1)) +#define vmv_v_x_u16m4(op0, op1) \ +__builtin_rvv_vmv_v_x_u16m4((uint16_t)(op0), (size_t)(op1)) +#define vmv_v_x_u16m8(op0, op1) \ +__builtin_rvv_vmv_v_x_u16m8((uint16_t)(op0), (size_t)(op1)) +#define vmv_v_x_u16mf2(op0, op1) \ +__builtin_rvv_vmv_v_x_u16mf2((uint16_t)(op0), (size_t)(op1)) +#define vmv_v_x_u16mf4(op0, op1) \ +__builtin_rvv_vmv_v_x_u16mf4((uint16_t)(op0), (size_t)(op1)) +#define vmv_v_x_u32m1(op0, op1) \ +__builtin_rvv_vmv_v_x_u32m1((uint32_t)(op0), (size_t)(op1)) +#define vmv_v_x_u32m2(op0, op1) \ +__builtin_rvv_vmv_v_x_u32m2((uint32_t)(op0), (size_t)(op1)) +#define vmv_v_x_u32m4(op0, op1) \ +__builtin_rvv_vmv_v_x_u32m4((uint32_t)(op0), (size_t)(op1)) +#define vmv_v_x_u32m8(op0, op1) \ +__builtin_rvv_vmv_v_x_u32m8((uint32_t)(op0), (size_t)(op1)) +#define vmv_v_x_u32mf2(op0, op1) \ +__builtin_rvv_vmv_v_x_u32mf2((uint32_t)(op0), (size_t)(op1)) +#define vmv_v_x_u64m1(op0, op1) \ +__builtin_rvv_vmv_v_x_u64m1((uint64_t)(op0), (size_t)(op1)) +#define vmv_v_x_u64m2(op0, op1) \ +__builtin_rvv_vmv_v_x_u64m2((uint64_t)(op0), (size_t)(op1)) +#define vmv_v_x_u64m4(op0, op1) \ +__builtin_rvv_vmv_v_x_u64m4((uint64_t)(op0), (size_t)(op1)) +#define vmv_v_x_u64m8(op0, op1) \ +__builtin_rvv_vmv_v_x_u64m8((uint64_t)(op0), (size_t)(op1)) +#define vsaddu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsadd_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vsadd_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsadd_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vsadd_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsadd_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vsadd_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsadd_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vsadd_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsadd_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vsadd_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsadd_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vsadd_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsadd_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vsadd_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsadd_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vsadd_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsadd_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vsadd_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsadd_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vsadd_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsadd_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsadd_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsadd_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsadd_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsadd_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsadd_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsadd_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsadd_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsadd_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsadd_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsadd_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsadd_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsadd_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsadd_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsadd_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsadd_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsadd_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsadd_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsadd_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsadd_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsadd_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsadd_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsadd_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsadd_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsadd_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsadd_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsadd_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsadd_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsadd_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vssubu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vssubu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vssubu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vssubu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vssubu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vssubu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vssubu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vssubu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vssubu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vssubu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vssubu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vssubu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vssubu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vssubu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vssubu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vssubu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vssubu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vssubu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vssubu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vssubu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vssubu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vssubu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vssubu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vssubu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vssubu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vssubu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vssubu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vssubu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vssubu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vssubu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vssubu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vssubu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vssubu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vssubu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vssubu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vssubu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vssubu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vssubu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vssubu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vssubu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vssubu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vssubu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vssubu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vssubu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vssubu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vssubu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vssubu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vssubu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vssubu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vssub_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vssub_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vssub_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vssub_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vssub_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vssub_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vssub_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vssub_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vssub_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vssub_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vssub_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vssub_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vssub_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vssub_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vssub_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vssub_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vssub_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vssub_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vssub_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vssub_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vssub_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vssub_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vssub_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vssub_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vssub_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vssub_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vssub_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vssub_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vssub_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vssub_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vssub_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vssub_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vssub_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vssub_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vssub_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vssub_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vssub_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vssub_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vssub_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vssub_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vssub_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vssub_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vssub_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vssub_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vssub_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vssub_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vssub_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vssub_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vssub_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vssub_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vssub_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vssub_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vssub_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vssub_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vssub_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vssub_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vssub_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vssub_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vssub_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vssub_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vssub_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vssub_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vssub_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vssub_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vssub_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vssub_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vssub_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vssub_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vssub_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vssub_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vssub_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vssub_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vssub_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vaadd_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vaadd_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vaadd_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vaadd_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vaadd_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vaadd_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vaadd_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vaadd_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vaadd_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vaadd_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vaadd_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vaadd_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vaadd_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vaadd_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vaadd_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vaadd_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vaadd_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vaadd_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vaadd_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vaadd_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vaadd_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vaadd_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vaadd_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vaadd_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vaadd_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vaadd_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vaadd_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vaadd_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vaadd_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vaadd_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vaadd_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vaadd_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vaadd_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vaadd_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vaadd_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vaadd_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vaadd_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vaadd_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vaadd_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vaadd_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vaadd_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vaadd_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vaadd_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vaadd_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vaadd_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vaadd_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vaadd_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vaadd_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vaadd_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vasubu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vasubu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vasubu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vasubu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vasubu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vasubu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vasubu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vasubu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vasubu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vasubu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vasubu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vasubu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vasubu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vasubu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vasubu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vasubu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vasubu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vasubu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vasubu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vasubu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vasubu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vasubu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vasubu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vasubu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vasubu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vasubu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vasubu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vasubu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vasubu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vasubu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vasubu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vasubu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vasubu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vasubu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vasubu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vasubu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vasubu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vasubu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vasubu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vasubu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vasubu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vasubu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vasubu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vasubu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vasubu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vasubu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vasubu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vasubu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vasubu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8m4((int8_t *)(op0), (vuint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8m8((int8_t *)(op0), (vuint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8m8_m((vbool1_t)(op0), (int8_t *)(op1), (vuint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vasub_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vasub_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vasub_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vasub_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vasub_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vasub_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vasub_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vasub_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vasub_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vasub_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vasub_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vasub_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vasub_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vasub_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vasub_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vasub_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vasub_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vasub_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vasub_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vasub_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vasub_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vasub_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vasub_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vasub_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vasub_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vasub_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vasub_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vasub_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vasub_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vasub_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vasub_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vasub_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vasub_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vasub_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vasub_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vasub_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vasub_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vasub_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vasub_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vasub_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vasub_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vasub_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vasub_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vasub_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vasub_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vasub_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vasub_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vasub_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vasub_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vasub_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vasub_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vasub_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vasub_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vasub_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vasub_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vasub_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vasub_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vasub_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vasub_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vasub_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vasub_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vasub_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vasub_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vasub_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vasub_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vasub_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vasub_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vasub_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vasub_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vasub_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vasub_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vasub_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vasub_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsmul_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vsmul_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsmul_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vsmul_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsmul_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vsmul_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsmul_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vsmul_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsmul_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vsmul_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsmul_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vsmul_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsmul_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vsmul_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsmul_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vsmul_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsmul_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vsmul_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsmul_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vsmul_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsmul_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsmul_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsmul_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsmul_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsmul_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsmul_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsmul_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsmul_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsmul_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsmul_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsmul_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsmul_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsmul_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsmul_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsmul_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsmul_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsmul_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsmul_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsmul_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsmul_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsmul_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsmul_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsmul_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsmul_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsmul_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsmul_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsmul_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsmul_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsmul_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vssrl_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vssrl_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vssrl_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vssrl_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vssrl_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vssrl_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vssrl_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vssrl_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vssrl_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vssrl_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vssrl_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vssrl_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vssrl_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vssrl_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vssrl_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vssrl_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vssrl_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vssrl_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vssrl_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vssrl_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8m1((vuint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8m2((vuint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8m4((vuint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8m8((vuint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8mf2((vuint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8mf4((vuint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8mf8((vuint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u16m1((vuint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u16m2((vuint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u16m4((vuint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u16m8((vuint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u16mf2((vuint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u16mf4((vuint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u32m1((vuint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u32m2((vuint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u32m4((vuint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u32m8((vuint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u32mf2((vuint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u64m1((vuint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u64m2((vuint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u64m4((vuint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u64m8((vuint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8m1((vint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vssra_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vssra_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8m2((vint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vssra_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vssra_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8m4((vint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vssra_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vssra_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8m8((vint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vssra_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vssra_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8mf2((vint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vssra_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vssra_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8mf4((vint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vssra_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vssra_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8mf8((vint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vssra_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vssra_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i16m1((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vssra_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vssra_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i16m2((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vssra_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vssra_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i16m4((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vssra_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vssra_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i16m8((vint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vssra_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vssra_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i16mf2((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vssra_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vssra_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i16mf4((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vssra_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vssra_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i32m1((vint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vssra_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vssra_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i32m2((vint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vssra_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vssra_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i32m4((vint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vssra_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vssra_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i32m8((vint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vssra_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vssra_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i32mf2((vint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vssra_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vssra_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i64m1((vint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vssra_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vssra_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i64m2((vint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vssra_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vssra_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i64m4((vint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vssra_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vssra_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i64m8((vint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vssra_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vssra_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8m1((vint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8m2((vint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8m4((vint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8m8((vint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8mf2((vint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8mf4((vint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8mf8((vint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i16m1((vint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i16m2((vint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i16m4((vint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i16m8((vint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i16mf2((vint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i16mf4((vint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i32m1((vint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i32m2((vint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i32m4((vint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i32m8((vint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i32mf2((vint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i64m1((vint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i64m2((vint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i64m4((vint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i64m8((vint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u8m1(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u8m1((vuint16m2_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u8m2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u8m2((vuint16m4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u8m4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u8m4((vuint16m8_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u8mf2((vuint16m1_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u8mf4((vuint16mf2_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u8mf8((vuint16mf4_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u16m1(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u16m1((vuint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u16m2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u16m2((vuint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u16m4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u16m4((vuint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u16mf2((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u16mf4((vuint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u32m1(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u32m1((vuint64m2_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u32m2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u32m2((vuint64m4_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u32m4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u32m4((vuint64m8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u32mf2((vuint64m1_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u8m1(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u8m1((vuint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u8m2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u8m2((vuint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u8m4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u8m4((vuint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u8mf2((vuint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u8mf4((vuint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u8mf8((vuint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u16m1(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u16m1((vuint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u16m2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u16m2((vuint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u16m4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u16m4((vuint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u16mf2((vuint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u16mf4((vuint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u32m1(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u32m1((vuint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u32m2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u32m2((vuint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u32m4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u32m4((vuint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u32mf2((vuint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnclip_wv_i8m1(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i8m1((vint16m2_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vnclip_wv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnclip_wv_i8m2(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i8m2((vint16m4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vnclip_wv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnclip_wv_i8m4(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i8m4((vint16m8_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vnclip_wv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnclip_wv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i8mf2((vint16m1_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vnclip_wv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnclip_wv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i8mf4((vint16mf2_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vnclip_wv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnclip_wv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i8mf8((vint16mf4_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vnclip_wv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnclip_wv_i16m1(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i16m1((vint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vnclip_wv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnclip_wv_i16m2(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i16m2((vint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vnclip_wv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnclip_wv_i16m4(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i16m4((vint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vnclip_wv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnclip_wv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i16mf2((vint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vnclip_wv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnclip_wv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i16mf4((vint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vnclip_wv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnclip_wv_i32m1(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i32m1((vint64m2_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vnclip_wv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnclip_wv_i32m2(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i32m2((vint64m4_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vnclip_wv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnclip_wv_i32m4(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i32m4((vint64m8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vnclip_wv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnclip_wv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i32mf2((vint64m1_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vnclip_wv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnclip_wx_i8m1(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i8m1((vint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i8m2(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i8m2((vint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i8m4(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i8m4((vint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i8mf2((vint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i8mf4((vint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i8mf8((vint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i16m1(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i16m1((vint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i16m2(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i16m2((vint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i16m4(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i16m4((vint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i16mf2((vint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i16mf4((vint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i32m1(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i32m1((vint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i32m2(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i32m2((vint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i32m4(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i32m4((vint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i32mf2((vint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8m1_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8m1_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8m1_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8m1_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8m2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8m2_i8m1((vint8m1_t)(op0), (vint8m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8m2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8m2_i8m1_m((vbool4_t)(op0), (vint8m1_t)(op1), (vint8m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8m4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8m4_i8m1((vint8m1_t)(op0), (vint8m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8m4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8m4_i8m1_m((vbool2_t)(op0), (vint8m1_t)(op1), (vint8m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8m8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8m8_i8m1((vint8m1_t)(op0), (vint8m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8m8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8m8_i8m1_m((vbool1_t)(op0), (vint8m1_t)(op1), (vint8m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8mf2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8mf2_i8m1((vint8m1_t)(op0), (vint8mf2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8mf2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8mf2_i8m1_m((vbool16_t)(op0), (vint8m1_t)(op1), (vint8mf2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8mf4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8mf4_i8m1((vint8m1_t)(op0), (vint8mf4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8mf4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8mf4_i8m1_m((vbool32_t)(op0), (vint8m1_t)(op1), (vint8mf4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8mf8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8mf8_i8m1((vint8m1_t)(op0), (vint8mf8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8mf8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8mf8_i8m1_m((vbool64_t)(op0), (vint8m1_t)(op1), (vint8mf8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i16m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i16m1_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i16m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i16m1_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i16m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i16m2_i16m1((vint16m1_t)(op0), (vint16m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i16m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i16m2_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint16m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i16m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i16m4_i16m1((vint16m1_t)(op0), (vint16m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i16m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i16m4_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint16m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i16m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i16m8_i16m1((vint16m1_t)(op0), (vint16m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i16m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i16m8_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint16m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i16mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i16mf2_i16m1((vint16m1_t)(op0), (vint16mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i16mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i16mf2_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint16mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i16mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i16mf4_i16m1((vint16m1_t)(op0), (vint16mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i16mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i16mf4_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint16mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i32m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i32m1_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i32m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i32m1_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i32m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i32m2_i32m1((vint32m1_t)(op0), (vint32m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i32m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i32m2_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint32m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i32m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i32m4_i32m1((vint32m1_t)(op0), (vint32m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i32m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i32m4_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint32m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i32m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i32m8_i32m1((vint32m1_t)(op0), (vint32m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i32m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i32m8_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint32m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i32mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i32mf2_i32m1((vint32m1_t)(op0), (vint32mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i32mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i32mf2_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint32mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i64m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i64m1_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i64m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i64m1_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i64m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i64m2_i64m1((vint64m1_t)(op0), (vint64m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i64m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i64m2_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint64m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i64m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i64m4_i64m1((vint64m1_t)(op0), (vint64m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i64m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i64m4_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint64m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i64m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i64m8_i64m1((vint64m1_t)(op0), (vint64m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i64m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i64m8_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint64m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8m1_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8m1_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8m1_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8m1_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8m2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8m2_u8m1((vuint8m1_t)(op0), (vuint8m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8m2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8m2_u8m1_m((vbool4_t)(op0), (vuint8m1_t)(op1), (vuint8m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8m4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8m4_u8m1((vuint8m1_t)(op0), (vuint8m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8m4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8m4_u8m1_m((vbool2_t)(op0), (vuint8m1_t)(op1), (vuint8m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8m8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8m8_u8m1((vuint8m1_t)(op0), (vuint8m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8m8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8m8_u8m1_m((vbool1_t)(op0), (vuint8m1_t)(op1), (vuint8m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8mf2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8mf2_u8m1((vuint8m1_t)(op0), (vuint8mf2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8mf2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8mf2_u8m1_m((vbool16_t)(op0), (vuint8m1_t)(op1), (vuint8mf2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8mf4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8mf4_u8m1((vuint8m1_t)(op0), (vuint8mf4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8mf4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8mf4_u8m1_m((vbool32_t)(op0), (vuint8m1_t)(op1), (vuint8mf4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8mf8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8mf8_u8m1((vuint8m1_t)(op0), (vuint8mf8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8mf8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8mf8_u8m1_m((vbool64_t)(op0), (vuint8m1_t)(op1), (vuint8mf8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u16m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u16m1_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u16m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u16m1_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u16m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u16m2_u16m1((vuint16m1_t)(op0), (vuint16m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u16m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u16m2_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint16m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u16m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u16m4_u16m1((vuint16m1_t)(op0), (vuint16m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u16m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u16m4_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint16m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u16m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u16m8_u16m1((vuint16m1_t)(op0), (vuint16m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u16m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u16m8_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint16m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u16mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u16mf2_u16m1((vuint16m1_t)(op0), (vuint16mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u16mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u16mf2_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint16mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u16mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u16mf4_u16m1((vuint16m1_t)(op0), (vuint16mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u16mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u16mf4_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint16mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u32m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u32m1_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u32m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u32m1_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u32m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u32m2_u32m1((vuint32m1_t)(op0), (vuint32m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u32m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u32m2_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint32m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u32m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u32m4_u32m1((vuint32m1_t)(op0), (vuint32m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u32m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u32m4_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint32m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u32m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u32m8_u32m1((vuint32m1_t)(op0), (vuint32m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u32m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u32m8_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint32m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u32mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u32mf2_u32m1((vuint32m1_t)(op0), (vuint32mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u32mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u32mf2_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint32mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u64m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u64m1_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u64m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u64m1_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u64m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u64m2_u64m1((vuint64m1_t)(op0), (vuint64m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u64m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u64m2_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint64m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u64m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u64m4_u64m1((vuint64m1_t)(op0), (vuint64m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u64m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u64m4_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint64m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u64m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u64m8_u64m1((vuint64m1_t)(op0), (vuint64m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u64m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u64m8_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint64m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8m1_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8m1_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8m1_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8m1_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8m2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8m2_u8m1((vuint8m1_t)(op0), (vuint8m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8m2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8m2_u8m1_m((vbool4_t)(op0), (vuint8m1_t)(op1), (vuint8m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8m4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8m4_u8m1((vuint8m1_t)(op0), (vuint8m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8m4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8m4_u8m1_m((vbool2_t)(op0), (vuint8m1_t)(op1), (vuint8m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8m8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8m8_u8m1((vuint8m1_t)(op0), (vuint8m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8m8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8m8_u8m1_m((vbool1_t)(op0), (vuint8m1_t)(op1), (vuint8m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8mf2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8mf2_u8m1((vuint8m1_t)(op0), (vuint8mf2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8mf2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8mf2_u8m1_m((vbool16_t)(op0), (vuint8m1_t)(op1), (vuint8mf2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8mf4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8mf4_u8m1((vuint8m1_t)(op0), (vuint8mf4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8mf4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8mf4_u8m1_m((vbool32_t)(op0), (vuint8m1_t)(op1), (vuint8mf4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8mf8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8mf8_u8m1((vuint8m1_t)(op0), (vuint8mf8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8mf8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8mf8_u8m1_m((vbool64_t)(op0), (vuint8m1_t)(op1), (vuint8mf8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u16m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u16m1_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u16m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u16m1_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u16m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u16m2_u16m1((vuint16m1_t)(op0), (vuint16m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u16m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u16m2_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint16m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u16m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u16m4_u16m1((vuint16m1_t)(op0), (vuint16m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u16m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u16m4_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint16m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u16m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u16m8_u16m1((vuint16m1_t)(op0), (vuint16m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u16m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u16m8_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint16m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u16mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u16mf2_u16m1((vuint16m1_t)(op0), (vuint16mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u16mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u16mf2_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint16mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u16mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u16mf4_u16m1((vuint16m1_t)(op0), (vuint16mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u16mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u16mf4_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint16mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u32m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u32m1_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u32m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u32m1_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u32m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u32m2_u32m1((vuint32m1_t)(op0), (vuint32m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u32m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u32m2_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint32m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u32m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u32m4_u32m1((vuint32m1_t)(op0), (vuint32m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u32m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u32m4_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint32m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u32m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u32m8_u32m1((vuint32m1_t)(op0), (vuint32m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u32m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u32m8_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint32m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u32mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u32mf2_u32m1((vuint32m1_t)(op0), (vuint32mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u32mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u32mf2_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint32mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u64m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u64m1_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u64m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u64m1_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u64m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u64m2_u64m1((vuint64m1_t)(op0), (vuint64m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u64m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u64m2_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint64m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u64m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u64m4_u64m1((vuint64m1_t)(op0), (vuint64m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u64m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u64m4_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint64m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u64m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u64m8_u64m1((vuint64m1_t)(op0), (vuint64m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u64m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u64m8_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint64m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8m1_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8m1_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8m1_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8m1_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8m2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8m2_i8m1((vint8m1_t)(op0), (vint8m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8m2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8m2_i8m1_m((vbool4_t)(op0), (vint8m1_t)(op1), (vint8m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8m4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8m4_i8m1((vint8m1_t)(op0), (vint8m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8m4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8m4_i8m1_m((vbool2_t)(op0), (vint8m1_t)(op1), (vint8m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8m8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8m8_i8m1((vint8m1_t)(op0), (vint8m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8m8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8m8_i8m1_m((vbool1_t)(op0), (vint8m1_t)(op1), (vint8m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8mf2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8mf2_i8m1((vint8m1_t)(op0), (vint8mf2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8mf2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8mf2_i8m1_m((vbool16_t)(op0), (vint8m1_t)(op1), (vint8mf2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8mf4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8mf4_i8m1((vint8m1_t)(op0), (vint8mf4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8mf4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8mf4_i8m1_m((vbool32_t)(op0), (vint8m1_t)(op1), (vint8mf4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8mf8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8mf8_i8m1((vint8m1_t)(op0), (vint8mf8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8mf8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8mf8_i8m1_m((vbool64_t)(op0), (vint8m1_t)(op1), (vint8mf8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i16m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i16m1_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i16m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i16m1_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i16m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i16m2_i16m1((vint16m1_t)(op0), (vint16m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i16m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i16m2_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint16m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i16m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i16m4_i16m1((vint16m1_t)(op0), (vint16m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i16m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i16m4_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint16m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i16m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i16m8_i16m1((vint16m1_t)(op0), (vint16m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i16m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i16m8_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint16m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i16mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i16mf2_i16m1((vint16m1_t)(op0), (vint16mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i16mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i16mf2_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint16mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i16mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i16mf4_i16m1((vint16m1_t)(op0), (vint16mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i16mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i16mf4_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint16mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i32m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i32m1_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i32m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i32m1_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i32m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i32m2_i32m1((vint32m1_t)(op0), (vint32m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i32m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i32m2_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint32m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i32m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i32m4_i32m1((vint32m1_t)(op0), (vint32m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i32m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i32m4_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint32m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i32m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i32m8_i32m1((vint32m1_t)(op0), (vint32m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i32m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i32m8_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint32m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i32mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i32mf2_i32m1((vint32m1_t)(op0), (vint32mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i32mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i32mf2_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint32mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i64m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i64m1_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i64m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i64m1_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i64m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i64m2_i64m1((vint64m1_t)(op0), (vint64m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i64m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i64m2_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint64m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i64m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i64m4_i64m1((vint64m1_t)(op0), (vint64m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i64m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i64m4_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint64m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i64m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i64m8_i64m1((vint64m1_t)(op0), (vint64m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i64m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i64m8_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint64m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8m1_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8m1_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8m1_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8m1_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8m2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8m2_u8m1((vuint8m1_t)(op0), (vuint8m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8m2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8m2_u8m1_m((vbool4_t)(op0), (vuint8m1_t)(op1), (vuint8m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8m4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8m4_u8m1((vuint8m1_t)(op0), (vuint8m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8m4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8m4_u8m1_m((vbool2_t)(op0), (vuint8m1_t)(op1), (vuint8m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8m8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8m8_u8m1((vuint8m1_t)(op0), (vuint8m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8m8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8m8_u8m1_m((vbool1_t)(op0), (vuint8m1_t)(op1), (vuint8m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8mf2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8mf2_u8m1((vuint8m1_t)(op0), (vuint8mf2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8mf2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8mf2_u8m1_m((vbool16_t)(op0), (vuint8m1_t)(op1), (vuint8mf2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8mf4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8mf4_u8m1((vuint8m1_t)(op0), (vuint8mf4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8mf4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8mf4_u8m1_m((vbool32_t)(op0), (vuint8m1_t)(op1), (vuint8mf4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8mf8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8mf8_u8m1((vuint8m1_t)(op0), (vuint8mf8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8mf8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8mf8_u8m1_m((vbool64_t)(op0), (vuint8m1_t)(op1), (vuint8mf8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u16m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u16m1_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u16m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u16m1_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u16m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u16m2_u16m1((vuint16m1_t)(op0), (vuint16m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u16m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u16m2_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint16m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u16m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u16m4_u16m1((vuint16m1_t)(op0), (vuint16m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u16m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u16m4_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint16m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u16m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u16m8_u16m1((vuint16m1_t)(op0), (vuint16m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u16m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u16m8_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint16m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u16mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u16mf2_u16m1((vuint16m1_t)(op0), (vuint16mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u16mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u16mf2_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint16mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u16mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u16mf4_u16m1((vuint16m1_t)(op0), (vuint16mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u16mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u16mf4_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint16mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u32m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u32m1_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u32m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u32m1_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u32m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u32m2_u32m1((vuint32m1_t)(op0), (vuint32m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u32m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u32m2_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint32m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u32m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u32m4_u32m1((vuint32m1_t)(op0), (vuint32m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u32m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u32m4_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint32m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u32m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u32m8_u32m1((vuint32m1_t)(op0), (vuint32m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u32m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u32m8_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint32m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u32mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u32mf2_u32m1((vuint32m1_t)(op0), (vuint32mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u32mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u32mf2_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint32mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u64m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u64m1_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u64m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u64m1_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u64m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u64m2_u64m1((vuint64m1_t)(op0), (vuint64m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u64m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u64m2_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint64m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u64m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u64m4_u64m1((vuint64m1_t)(op0), (vuint64m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u64m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u64m4_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint64m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u64m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u64m8_u64m1((vuint64m1_t)(op0), (vuint64m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u64m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u64m8_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint64m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8m1_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8m1_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8m1_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8m1_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8m2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8m2_i8m1((vint8m1_t)(op0), (vint8m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8m2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8m2_i8m1_m((vbool4_t)(op0), (vint8m1_t)(op1), (vint8m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8m4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8m4_i8m1((vint8m1_t)(op0), (vint8m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8m4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8m4_i8m1_m((vbool2_t)(op0), (vint8m1_t)(op1), (vint8m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8m8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8m8_i8m1((vint8m1_t)(op0), (vint8m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8m8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8m8_i8m1_m((vbool1_t)(op0), (vint8m1_t)(op1), (vint8m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8mf2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8mf2_i8m1((vint8m1_t)(op0), (vint8mf2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8mf2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8mf2_i8m1_m((vbool16_t)(op0), (vint8m1_t)(op1), (vint8mf2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8mf4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8mf4_i8m1((vint8m1_t)(op0), (vint8mf4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8mf4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8mf4_i8m1_m((vbool32_t)(op0), (vint8m1_t)(op1), (vint8mf4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8mf8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8mf8_i8m1((vint8m1_t)(op0), (vint8mf8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8mf8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8mf8_i8m1_m((vbool64_t)(op0), (vint8m1_t)(op1), (vint8mf8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i16m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i16m1_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i16m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i16m1_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i16m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i16m2_i16m1((vint16m1_t)(op0), (vint16m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i16m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i16m2_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint16m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i16m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i16m4_i16m1((vint16m1_t)(op0), (vint16m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i16m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i16m4_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint16m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i16m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i16m8_i16m1((vint16m1_t)(op0), (vint16m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i16m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i16m8_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint16m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i16mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i16mf2_i16m1((vint16m1_t)(op0), (vint16mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i16mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i16mf2_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint16mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i16mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i16mf4_i16m1((vint16m1_t)(op0), (vint16mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i16mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i16mf4_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint16mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i32m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i32m1_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i32m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i32m1_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i32m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i32m2_i32m1((vint32m1_t)(op0), (vint32m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i32m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i32m2_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint32m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i32m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i32m4_i32m1((vint32m1_t)(op0), (vint32m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i32m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i32m4_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint32m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i32m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i32m8_i32m1((vint32m1_t)(op0), (vint32m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i32m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i32m8_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint32m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i32mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i32mf2_i32m1((vint32m1_t)(op0), (vint32mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i32mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i32mf2_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint32mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i64m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i64m1_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i64m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i64m1_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i64m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i64m2_i64m1((vint64m1_t)(op0), (vint64m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i64m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i64m2_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint64m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i64m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i64m4_i64m1((vint64m1_t)(op0), (vint64m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i64m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i64m4_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint64m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i64m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i64m8_i64m1((vint64m1_t)(op0), (vint64m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i64m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i64m8_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint64m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8m1_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8m1_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8m1_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8m1_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8m2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8m2_i8m1((vint8m1_t)(op0), (vint8m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8m2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8m2_i8m1_m((vbool4_t)(op0), (vint8m1_t)(op1), (vint8m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8m4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8m4_i8m1((vint8m1_t)(op0), (vint8m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8m4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8m4_i8m1_m((vbool2_t)(op0), (vint8m1_t)(op1), (vint8m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8m8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8m8_i8m1((vint8m1_t)(op0), (vint8m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8m8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8m8_i8m1_m((vbool1_t)(op0), (vint8m1_t)(op1), (vint8m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8mf2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8mf2_i8m1((vint8m1_t)(op0), (vint8mf2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8mf2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8mf2_i8m1_m((vbool16_t)(op0), (vint8m1_t)(op1), (vint8mf2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8mf4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8mf4_i8m1((vint8m1_t)(op0), (vint8mf4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8mf4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8mf4_i8m1_m((vbool32_t)(op0), (vint8m1_t)(op1), (vint8mf4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8mf8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8mf8_i8m1((vint8m1_t)(op0), (vint8mf8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8mf8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8mf8_i8m1_m((vbool64_t)(op0), (vint8m1_t)(op1), (vint8mf8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i16m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i16m1_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i16m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i16m1_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i16m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i16m2_i16m1((vint16m1_t)(op0), (vint16m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i16m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i16m2_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint16m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i16m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i16m4_i16m1((vint16m1_t)(op0), (vint16m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i16m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i16m4_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint16m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i16m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i16m8_i16m1((vint16m1_t)(op0), (vint16m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i16m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i16m8_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint16m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i16mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i16mf2_i16m1((vint16m1_t)(op0), (vint16mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i16mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i16mf2_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint16mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i16mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i16mf4_i16m1((vint16m1_t)(op0), (vint16mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i16mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i16mf4_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint16mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i32m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i32m1_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i32m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i32m1_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i32m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i32m2_i32m1((vint32m1_t)(op0), (vint32m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i32m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i32m2_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint32m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i32m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i32m4_i32m1((vint32m1_t)(op0), (vint32m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i32m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i32m4_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint32m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i32m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i32m8_i32m1((vint32m1_t)(op0), (vint32m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i32m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i32m8_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint32m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i32mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i32mf2_i32m1((vint32m1_t)(op0), (vint32mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i32mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i32mf2_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint32mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i64m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i64m1_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i64m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i64m1_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i64m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i64m2_i64m1((vint64m1_t)(op0), (vint64m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i64m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i64m2_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint64m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i64m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i64m4_i64m1((vint64m1_t)(op0), (vint64m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i64m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i64m4_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint64m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i64m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i64m8_i64m1((vint64m1_t)(op0), (vint64m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i64m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i64m8_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint64m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8m1_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8m1_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8m1_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8m1_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8m2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8m2_u8m1((vuint8m1_t)(op0), (vuint8m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8m2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8m2_u8m1_m((vbool4_t)(op0), (vuint8m1_t)(op1), (vuint8m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8m4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8m4_u8m1((vuint8m1_t)(op0), (vuint8m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8m4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8m4_u8m1_m((vbool2_t)(op0), (vuint8m1_t)(op1), (vuint8m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8m8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8m8_u8m1((vuint8m1_t)(op0), (vuint8m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8m8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8m8_u8m1_m((vbool1_t)(op0), (vuint8m1_t)(op1), (vuint8m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8mf2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8mf2_u8m1((vuint8m1_t)(op0), (vuint8mf2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8mf2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8mf2_u8m1_m((vbool16_t)(op0), (vuint8m1_t)(op1), (vuint8mf2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8mf4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8mf4_u8m1((vuint8m1_t)(op0), (vuint8mf4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8mf4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8mf4_u8m1_m((vbool32_t)(op0), (vuint8m1_t)(op1), (vuint8mf4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8mf8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8mf8_u8m1((vuint8m1_t)(op0), (vuint8mf8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8mf8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8mf8_u8m1_m((vbool64_t)(op0), (vuint8m1_t)(op1), (vuint8mf8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u16m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u16m1_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u16m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u16m1_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u16m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u16m2_u16m1((vuint16m1_t)(op0), (vuint16m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u16m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u16m2_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint16m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u16m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u16m4_u16m1((vuint16m1_t)(op0), (vuint16m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u16m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u16m4_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint16m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u16m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u16m8_u16m1((vuint16m1_t)(op0), (vuint16m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u16m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u16m8_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint16m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u16mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u16mf2_u16m1((vuint16m1_t)(op0), (vuint16mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u16mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u16mf2_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint16mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u16mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u16mf4_u16m1((vuint16m1_t)(op0), (vuint16mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u16mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u16mf4_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint16mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u32m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u32m1_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u32m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u32m1_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u32m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u32m2_u32m1((vuint32m1_t)(op0), (vuint32m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u32m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u32m2_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint32m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u32m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u32m4_u32m1((vuint32m1_t)(op0), (vuint32m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u32m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u32m4_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint32m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u32m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u32m8_u32m1((vuint32m1_t)(op0), (vuint32m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u32m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u32m8_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint32m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u32mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u32mf2_u32m1((vuint32m1_t)(op0), (vuint32mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u32mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u32mf2_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint32mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u64m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u64m1_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u64m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u64m1_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u64m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u64m2_u64m1((vuint64m1_t)(op0), (vuint64m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u64m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u64m2_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint64m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u64m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u64m4_u64m1((vuint64m1_t)(op0), (vuint64m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u64m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u64m4_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint64m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u64m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u64m8_u64m1((vuint64m1_t)(op0), (vuint64m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u64m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u64m8_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint64m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vredor_vs_i8m1_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8m1_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8m1_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8m1_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i8m2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8m2_i8m1((vint8m1_t)(op0), (vint8m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8m2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8m2_i8m1_m((vbool4_t)(op0), (vint8m1_t)(op1), (vint8m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i8m4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8m4_i8m1((vint8m1_t)(op0), (vint8m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8m4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8m4_i8m1_m((vbool2_t)(op0), (vint8m1_t)(op1), (vint8m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i8m8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8m8_i8m1((vint8m1_t)(op0), (vint8m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8m8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8m8_i8m1_m((vbool1_t)(op0), (vint8m1_t)(op1), (vint8m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i8mf2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8mf2_i8m1((vint8m1_t)(op0), (vint8mf2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8mf2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8mf2_i8m1_m((vbool16_t)(op0), (vint8m1_t)(op1), (vint8mf2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i8mf4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8mf4_i8m1((vint8m1_t)(op0), (vint8mf4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8mf4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8mf4_i8m1_m((vbool32_t)(op0), (vint8m1_t)(op1), (vint8mf4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i8mf8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8mf8_i8m1((vint8m1_t)(op0), (vint8mf8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8mf8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8mf8_i8m1_m((vbool64_t)(op0), (vint8m1_t)(op1), (vint8mf8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i16m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i16m1_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i16m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i16m1_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i16m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i16m2_i16m1((vint16m1_t)(op0), (vint16m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i16m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i16m2_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint16m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i16m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i16m4_i16m1((vint16m1_t)(op0), (vint16m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i16m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i16m4_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint16m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i16m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i16m8_i16m1((vint16m1_t)(op0), (vint16m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i16m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i16m8_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint16m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i16mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i16mf2_i16m1((vint16m1_t)(op0), (vint16mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i16mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i16mf2_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint16mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i16mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i16mf4_i16m1((vint16m1_t)(op0), (vint16mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i16mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i16mf4_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint16mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i32m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i32m1_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i32m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i32m1_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i32m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i32m2_i32m1((vint32m1_t)(op0), (vint32m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i32m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i32m2_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint32m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i32m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i32m4_i32m1((vint32m1_t)(op0), (vint32m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i32m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i32m4_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint32m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i32m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i32m8_i32m1((vint32m1_t)(op0), (vint32m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i32m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i32m8_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint32m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i32mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i32mf2_i32m1((vint32m1_t)(op0), (vint32mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i32mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i32mf2_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint32mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i64m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i64m1_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i64m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i64m1_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i64m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i64m2_i64m1((vint64m1_t)(op0), (vint64m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i64m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i64m2_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint64m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i64m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i64m4_i64m1((vint64m1_t)(op0), (vint64m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i64m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i64m4_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint64m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i64m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i64m8_i64m1((vint64m1_t)(op0), (vint64m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i64m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i64m8_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint64m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8m1_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8m1_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8m1_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8m1_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8m2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8m2_u8m1((vuint8m1_t)(op0), (vuint8m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8m2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8m2_u8m1_m((vbool4_t)(op0), (vuint8m1_t)(op1), (vuint8m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8m4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8m4_u8m1((vuint8m1_t)(op0), (vuint8m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8m4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8m4_u8m1_m((vbool2_t)(op0), (vuint8m1_t)(op1), (vuint8m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8m8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8m8_u8m1((vuint8m1_t)(op0), (vuint8m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8m8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8m8_u8m1_m((vbool1_t)(op0), (vuint8m1_t)(op1), (vuint8m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8mf2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8mf2_u8m1((vuint8m1_t)(op0), (vuint8mf2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8mf2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8mf2_u8m1_m((vbool16_t)(op0), (vuint8m1_t)(op1), (vuint8mf2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8mf4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8mf4_u8m1((vuint8m1_t)(op0), (vuint8mf4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8mf4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8mf4_u8m1_m((vbool32_t)(op0), (vuint8m1_t)(op1), (vuint8mf4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8mf8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8mf8_u8m1((vuint8m1_t)(op0), (vuint8mf8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8mf8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8mf8_u8m1_m((vbool64_t)(op0), (vuint8m1_t)(op1), (vuint8mf8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u16m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u16m1_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u16m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u16m1_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u16m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u16m2_u16m1((vuint16m1_t)(op0), (vuint16m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u16m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u16m2_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint16m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u16m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u16m4_u16m1((vuint16m1_t)(op0), (vuint16m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u16m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u16m4_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint16m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u16m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u16m8_u16m1((vuint16m1_t)(op0), (vuint16m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u16m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u16m8_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint16m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u16mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u16mf2_u16m1((vuint16m1_t)(op0), (vuint16mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u16mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u16mf2_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint16mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u16mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u16mf4_u16m1((vuint16m1_t)(op0), (vuint16mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u16mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u16mf4_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint16mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u32m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u32m1_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u32m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u32m1_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u32m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u32m2_u32m1((vuint32m1_t)(op0), (vuint32m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u32m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u32m2_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint32m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u32m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u32m4_u32m1((vuint32m1_t)(op0), (vuint32m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u32m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u32m4_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint32m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u32m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u32m8_u32m1((vuint32m1_t)(op0), (vuint32m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u32m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u32m8_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint32m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u32mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u32mf2_u32m1((vuint32m1_t)(op0), (vuint32mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u32mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u32mf2_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint32mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u64m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u64m1_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u64m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u64m1_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u64m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u64m2_u64m1((vuint64m1_t)(op0), (vuint64m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u64m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u64m2_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint64m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u64m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u64m4_u64m1((vuint64m1_t)(op0), (vuint64m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u64m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u64m4_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint64m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u64m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u64m8_u64m1((vuint64m1_t)(op0), (vuint64m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u64m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u64m8_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint64m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8m1_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8m1_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8m1_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8m1_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8m2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8m2_i8m1((vint8m1_t)(op0), (vint8m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8m2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8m2_i8m1_m((vbool4_t)(op0), (vint8m1_t)(op1), (vint8m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8m4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8m4_i8m1((vint8m1_t)(op0), (vint8m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8m4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8m4_i8m1_m((vbool2_t)(op0), (vint8m1_t)(op1), (vint8m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8m8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8m8_i8m1((vint8m1_t)(op0), (vint8m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8m8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8m8_i8m1_m((vbool1_t)(op0), (vint8m1_t)(op1), (vint8m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8mf2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8mf2_i8m1((vint8m1_t)(op0), (vint8mf2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8mf2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8mf2_i8m1_m((vbool16_t)(op0), (vint8m1_t)(op1), (vint8mf2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8mf4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8mf4_i8m1((vint8m1_t)(op0), (vint8mf4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8mf4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8mf4_i8m1_m((vbool32_t)(op0), (vint8m1_t)(op1), (vint8mf4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8mf8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8mf8_i8m1((vint8m1_t)(op0), (vint8mf8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8mf8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8mf8_i8m1_m((vbool64_t)(op0), (vint8m1_t)(op1), (vint8mf8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i16m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i16m1_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i16m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i16m1_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i16m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i16m2_i16m1((vint16m1_t)(op0), (vint16m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i16m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i16m2_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint16m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i16m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i16m4_i16m1((vint16m1_t)(op0), (vint16m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i16m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i16m4_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint16m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i16m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i16m8_i16m1((vint16m1_t)(op0), (vint16m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i16m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i16m8_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint16m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i16mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i16mf2_i16m1((vint16m1_t)(op0), (vint16mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i16mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i16mf2_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint16mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i16mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i16mf4_i16m1((vint16m1_t)(op0), (vint16mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i16mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i16mf4_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint16mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i32m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i32m1_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i32m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i32m1_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i32m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i32m2_i32m1((vint32m1_t)(op0), (vint32m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i32m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i32m2_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint32m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i32m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i32m4_i32m1((vint32m1_t)(op0), (vint32m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i32m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i32m4_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint32m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i32m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i32m8_i32m1((vint32m1_t)(op0), (vint32m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i32m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i32m8_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint32m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i32mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i32mf2_i32m1((vint32m1_t)(op0), (vint32mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i32mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i32mf2_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint32mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i64m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i64m1_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i64m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i64m1_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i64m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i64m2_i64m1((vint64m1_t)(op0), (vint64m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i64m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i64m2_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint64m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i64m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i64m4_i64m1((vint64m1_t)(op0), (vint64m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i64m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i64m4_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint64m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i64m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i64m8_i64m1((vint64m1_t)(op0), (vint64m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i64m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i64m8_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint64m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8m1_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8m1_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8m1_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8m1_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8m2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8m2_u8m1((vuint8m1_t)(op0), (vuint8m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8m2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8m2_u8m1_m((vbool4_t)(op0), (vuint8m1_t)(op1), (vuint8m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8m4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8m4_u8m1((vuint8m1_t)(op0), (vuint8m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8m4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8m4_u8m1_m((vbool2_t)(op0), (vuint8m1_t)(op1), (vuint8m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8m8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8m8_u8m1((vuint8m1_t)(op0), (vuint8m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8m8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8m8_u8m1_m((vbool1_t)(op0), (vuint8m1_t)(op1), (vuint8m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8mf2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8mf2_u8m1((vuint8m1_t)(op0), (vuint8mf2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8mf2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8mf2_u8m1_m((vbool16_t)(op0), (vuint8m1_t)(op1), (vuint8mf2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8mf4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8mf4_u8m1((vuint8m1_t)(op0), (vuint8mf4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8mf4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8mf4_u8m1_m((vbool32_t)(op0), (vuint8m1_t)(op1), (vuint8mf4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8mf8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8mf8_u8m1((vuint8m1_t)(op0), (vuint8mf8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8mf8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8mf8_u8m1_m((vbool64_t)(op0), (vuint8m1_t)(op1), (vuint8mf8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u16m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u16m1_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u16m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u16m1_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u16m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u16m2_u16m1((vuint16m1_t)(op0), (vuint16m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u16m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u16m2_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint16m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u16m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u16m4_u16m1((vuint16m1_t)(op0), (vuint16m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u16m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u16m4_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint16m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u16m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u16m8_u16m1((vuint16m1_t)(op0), (vuint16m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u16m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u16m8_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint16m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u16mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u16mf2_u16m1((vuint16m1_t)(op0), (vuint16mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u16mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u16mf2_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint16mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u16mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u16mf4_u16m1((vuint16m1_t)(op0), (vuint16mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u16mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u16mf4_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint16mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u32m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u32m1_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u32m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u32m1_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u32m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u32m2_u32m1((vuint32m1_t)(op0), (vuint32m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u32m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u32m2_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint32m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u32m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u32m4_u32m1((vuint32m1_t)(op0), (vuint32m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u32m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u32m4_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint32m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u32m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u32m8_u32m1((vuint32m1_t)(op0), (vuint32m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u32m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u32m8_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint32m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u32mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u32mf2_u32m1((vuint32m1_t)(op0), (vuint32mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u32mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u32mf2_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint32mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u64m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u64m1_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u64m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u64m1_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u64m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u64m2_u64m1((vuint64m1_t)(op0), (vuint64m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u64m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u64m2_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint64m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u64m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u64m4_u64m1((vuint64m1_t)(op0), (vuint64m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u64m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u64m4_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint64m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u64m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u64m8_u64m1((vuint64m1_t)(op0), (vuint64m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u64m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u64m8_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint64m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8m1_i16m1((vint16m1_t)(op0), (vint8m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8m1_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint8m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8m2_i16m1((vint16m1_t)(op0), (vint8m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8m2_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint8m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8m4_i16m1((vint16m1_t)(op0), (vint8m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8m4_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint8m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8m8_i16m1((vint16m1_t)(op0), (vint8m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8m8_i16m1_m((vbool1_t)(op0), (vint16m1_t)(op1), (vint8m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8mf2_i16m1((vint16m1_t)(op0), (vint8mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8mf2_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8mf4_i16m1((vint16m1_t)(op0), (vint8mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8mf4_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint8mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8mf8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8mf8_i16m1((vint16m1_t)(op0), (vint8mf8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8mf8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8mf8_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint8mf8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i16m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i16m1_i32m1((vint32m1_t)(op0), (vint16m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i16m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i16m1_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint16m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i16m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i16m2_i32m1((vint32m1_t)(op0), (vint16m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i16m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i16m2_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint16m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i16m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i16m4_i32m1((vint32m1_t)(op0), (vint16m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i16m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i16m4_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint16m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i16m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i16m8_i32m1((vint32m1_t)(op0), (vint16m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i16m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i16m8_i32m1_m((vbool2_t)(op0), (vint32m1_t)(op1), (vint16m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i16mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i16mf2_i32m1((vint32m1_t)(op0), (vint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i16mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i16mf2_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i16mf4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i16mf4_i32m1((vint32m1_t)(op0), (vint16mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i16mf4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i16mf4_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint16mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i32m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i32m1_i64m1((vint64m1_t)(op0), (vint32m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i32m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i32m1_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint32m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i32m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i32m2_i64m1((vint64m1_t)(op0), (vint32m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i32m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i32m2_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint32m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i32m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i32m4_i64m1((vint64m1_t)(op0), (vint32m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i32m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i32m4_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint32m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i32m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i32m8_i64m1((vint64m1_t)(op0), (vint32m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i32m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i32m8_i64m1_m((vbool4_t)(op0), (vint64m1_t)(op1), (vint32m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i32mf2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i32mf2_i64m1((vint64m1_t)(op0), (vint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i32mf2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i32mf2_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8m1_u16m1((vuint16m1_t)(op0), (vuint8m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8m1_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint8m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8m2_u16m1((vuint16m1_t)(op0), (vuint8m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8m2_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint8m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8m4_u16m1((vuint16m1_t)(op0), (vuint8m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8m4_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint8m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8m8_u16m1((vuint16m1_t)(op0), (vuint8m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8m8_u16m1_m((vbool1_t)(op0), (vuint16m1_t)(op1), (vuint8m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8mf2_u16m1((vuint16m1_t)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8mf2_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8mf4_u16m1((vuint16m1_t)(op0), (vuint8mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8mf4_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint8mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8mf8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8mf8_u16m1((vuint16m1_t)(op0), (vuint8mf8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8mf8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8mf8_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint8mf8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u16m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u16m1_u32m1((vuint32m1_t)(op0), (vuint16m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u16m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u16m1_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint16m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u16m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u16m2_u32m1((vuint32m1_t)(op0), (vuint16m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u16m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u16m2_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint16m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u16m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u16m4_u32m1((vuint32m1_t)(op0), (vuint16m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u16m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u16m4_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint16m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u16m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u16m8_u32m1((vuint32m1_t)(op0), (vuint16m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u16m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u16m8_u32m1_m((vbool2_t)(op0), (vuint32m1_t)(op1), (vuint16m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u16mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u16mf2_u32m1((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u16mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u16mf2_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u16mf4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u16mf4_u32m1((vuint32m1_t)(op0), (vuint16mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u16mf4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u16mf4_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint16mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u32m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u32m1_u64m1((vuint64m1_t)(op0), (vuint32m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u32m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u32m1_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint32m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u32m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u32m2_u64m1((vuint64m1_t)(op0), (vuint32m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u32m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u32m2_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint32m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u32m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u32m4_u64m1((vuint64m1_t)(op0), (vuint32m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u32m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u32m4_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint32m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u32m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u32m8_u64m1((vuint64m1_t)(op0), (vuint32m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u32m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u32m8_u64m1_m((vbool4_t)(op0), (vuint64m1_t)(op1), (vuint32m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u32mf2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u32mf2_u64m1((vuint64m1_t)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u32mf2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u32mf2_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8m4((uint8_t *)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8m8((uint8_t *)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8m8_m((vbool1_t)(op0), (uint8_t *)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmnot_m_b8(op0, op1) \ +__builtin_rvv_vmnot_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vmnot_m_b4(op0, op1) \ +__builtin_rvv_vmnot_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vmnot_m_b2(op0, op1) \ +__builtin_rvv_vmnot_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vmnot_m_b1(op0, op1) \ +__builtin_rvv_vmnot_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vmnot_m_b16(op0, op1) \ +__builtin_rvv_vmnot_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vmnot_m_b32(op0, op1) \ +__builtin_rvv_vmnot_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vmnot_m_b64(op0, op1) \ +__builtin_rvv_vmnot_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define viota_m_u8m1(op0, op1) \ +__builtin_rvv_viota_m_u8m1((vbool8_t)(op0), (size_t)(op1)) +#define viota_m_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define viota_m_u8m2(op0, op1) \ +__builtin_rvv_viota_m_u8m2((vbool4_t)(op0), (size_t)(op1)) +#define viota_m_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define viota_m_u8m4(op0, op1) \ +__builtin_rvv_viota_m_u8m4((vbool2_t)(op0), (size_t)(op1)) +#define viota_m_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define viota_m_u8m8(op0, op1) \ +__builtin_rvv_viota_m_u8m8((vbool1_t)(op0), (size_t)(op1)) +#define viota_m_u8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define viota_m_u8mf2(op0, op1) \ +__builtin_rvv_viota_m_u8mf2((vbool16_t)(op0), (size_t)(op1)) +#define viota_m_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define viota_m_u8mf4(op0, op1) \ +__builtin_rvv_viota_m_u8mf4((vbool32_t)(op0), (size_t)(op1)) +#define viota_m_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define viota_m_u8mf8(op0, op1) \ +__builtin_rvv_viota_m_u8mf8((vbool64_t)(op0), (size_t)(op1)) +#define viota_m_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define viota_m_u16m1(op0, op1) \ +__builtin_rvv_viota_m_u16m1((vbool16_t)(op0), (size_t)(op1)) +#define viota_m_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define viota_m_u16m2(op0, op1) \ +__builtin_rvv_viota_m_u16m2((vbool8_t)(op0), (size_t)(op1)) +#define viota_m_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define viota_m_u16m4(op0, op1) \ +__builtin_rvv_viota_m_u16m4((vbool4_t)(op0), (size_t)(op1)) +#define viota_m_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define viota_m_u16m8(op0, op1) \ +__builtin_rvv_viota_m_u16m8((vbool2_t)(op0), (size_t)(op1)) +#define viota_m_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define viota_m_u16mf2(op0, op1) \ +__builtin_rvv_viota_m_u16mf2((vbool32_t)(op0), (size_t)(op1)) +#define viota_m_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define viota_m_u16mf4(op0, op1) \ +__builtin_rvv_viota_m_u16mf4((vbool64_t)(op0), (size_t)(op1)) +#define viota_m_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define viota_m_u32m1(op0, op1) \ +__builtin_rvv_viota_m_u32m1((vbool32_t)(op0), (size_t)(op1)) +#define viota_m_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define viota_m_u32m2(op0, op1) \ +__builtin_rvv_viota_m_u32m2((vbool16_t)(op0), (size_t)(op1)) +#define viota_m_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define viota_m_u32m4(op0, op1) \ +__builtin_rvv_viota_m_u32m4((vbool8_t)(op0), (size_t)(op1)) +#define viota_m_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define viota_m_u32m8(op0, op1) \ +__builtin_rvv_viota_m_u32m8((vbool4_t)(op0), (size_t)(op1)) +#define viota_m_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define viota_m_u32mf2(op0, op1) \ +__builtin_rvv_viota_m_u32mf2((vbool64_t)(op0), (size_t)(op1)) +#define viota_m_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define viota_m_u64m1(op0, op1) \ +__builtin_rvv_viota_m_u64m1((vbool64_t)(op0), (size_t)(op1)) +#define viota_m_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define viota_m_u64m2(op0, op1) \ +__builtin_rvv_viota_m_u64m2((vbool32_t)(op0), (size_t)(op1)) +#define viota_m_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define viota_m_u64m4(op0, op1) \ +__builtin_rvv_viota_m_u64m4((vbool16_t)(op0), (size_t)(op1)) +#define viota_m_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define viota_m_u64m8(op0, op1) \ +__builtin_rvv_viota_m_u64m8((vbool8_t)(op0), (size_t)(op1)) +#define viota_m_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vid_v_i8m1(op0) \ +__builtin_rvv_vid_v_i8m1((size_t)(op0)) +#define vid_v_i8m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vid_v_i8m2(op0) \ +__builtin_rvv_vid_v_i8m2((size_t)(op0)) +#define vid_v_i8m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vid_v_i8m4(op0) \ +__builtin_rvv_vid_v_i8m4((size_t)(op0)) +#define vid_v_i8m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vid_v_i8m8(op0) \ +__builtin_rvv_vid_v_i8m8((size_t)(op0)) +#define vid_v_i8m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vid_v_i8mf2(op0) \ +__builtin_rvv_vid_v_i8mf2((size_t)(op0)) +#define vid_v_i8mf2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vid_v_i8mf4(op0) \ +__builtin_rvv_vid_v_i8mf4((size_t)(op0)) +#define vid_v_i8mf4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vid_v_i8mf8(op0) \ +__builtin_rvv_vid_v_i8mf8((size_t)(op0)) +#define vid_v_i8mf8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vid_v_i16m1(op0) \ +__builtin_rvv_vid_v_i16m1((size_t)(op0)) +#define vid_v_i16m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vid_v_i16m2(op0) \ +__builtin_rvv_vid_v_i16m2((size_t)(op0)) +#define vid_v_i16m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vid_v_i16m4(op0) \ +__builtin_rvv_vid_v_i16m4((size_t)(op0)) +#define vid_v_i16m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vid_v_i16m8(op0) \ +__builtin_rvv_vid_v_i16m8((size_t)(op0)) +#define vid_v_i16m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vid_v_i16mf2(op0) \ +__builtin_rvv_vid_v_i16mf2((size_t)(op0)) +#define vid_v_i16mf2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vid_v_i16mf4(op0) \ +__builtin_rvv_vid_v_i16mf4((size_t)(op0)) +#define vid_v_i16mf4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vid_v_i32m1(op0) \ +__builtin_rvv_vid_v_i32m1((size_t)(op0)) +#define vid_v_i32m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vid_v_i32m2(op0) \ +__builtin_rvv_vid_v_i32m2((size_t)(op0)) +#define vid_v_i32m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vid_v_i32m4(op0) \ +__builtin_rvv_vid_v_i32m4((size_t)(op0)) +#define vid_v_i32m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vid_v_i32m8(op0) \ +__builtin_rvv_vid_v_i32m8((size_t)(op0)) +#define vid_v_i32m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vid_v_i32mf2(op0) \ +__builtin_rvv_vid_v_i32mf2((size_t)(op0)) +#define vid_v_i32mf2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vid_v_i64m1(op0) \ +__builtin_rvv_vid_v_i64m1((size_t)(op0)) +#define vid_v_i64m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vid_v_i64m2(op0) \ +__builtin_rvv_vid_v_i64m2((size_t)(op0)) +#define vid_v_i64m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vid_v_i64m4(op0) \ +__builtin_rvv_vid_v_i64m4((size_t)(op0)) +#define vid_v_i64m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vid_v_i64m8(op0) \ +__builtin_rvv_vid_v_i64m8((size_t)(op0)) +#define vid_v_i64m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vid_v_u8m1(op0) \ +__builtin_rvv_vid_v_u8m1((size_t)(op0)) +#define vid_v_u8m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vid_v_u8m2(op0) \ +__builtin_rvv_vid_v_u8m2((size_t)(op0)) +#define vid_v_u8m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vid_v_u8m4(op0) \ +__builtin_rvv_vid_v_u8m4((size_t)(op0)) +#define vid_v_u8m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vid_v_u8m8(op0) \ +__builtin_rvv_vid_v_u8m8((size_t)(op0)) +#define vid_v_u8m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vid_v_u8mf2(op0) \ +__builtin_rvv_vid_v_u8mf2((size_t)(op0)) +#define vid_v_u8mf2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vid_v_u8mf4(op0) \ +__builtin_rvv_vid_v_u8mf4((size_t)(op0)) +#define vid_v_u8mf4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vid_v_u8mf8(op0) \ +__builtin_rvv_vid_v_u8mf8((size_t)(op0)) +#define vid_v_u8mf8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vid_v_u16m1(op0) \ +__builtin_rvv_vid_v_u16m1((size_t)(op0)) +#define vid_v_u16m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vid_v_u16m2(op0) \ +__builtin_rvv_vid_v_u16m2((size_t)(op0)) +#define vid_v_u16m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vid_v_u16m4(op0) \ +__builtin_rvv_vid_v_u16m4((size_t)(op0)) +#define vid_v_u16m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vid_v_u16m8(op0) \ +__builtin_rvv_vid_v_u16m8((size_t)(op0)) +#define vid_v_u16m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vid_v_u16mf2(op0) \ +__builtin_rvv_vid_v_u16mf2((size_t)(op0)) +#define vid_v_u16mf2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vid_v_u16mf4(op0) \ +__builtin_rvv_vid_v_u16mf4((size_t)(op0)) +#define vid_v_u16mf4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vid_v_u32m1(op0) \ +__builtin_rvv_vid_v_u32m1((size_t)(op0)) +#define vid_v_u32m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vid_v_u32m2(op0) \ +__builtin_rvv_vid_v_u32m2((size_t)(op0)) +#define vid_v_u32m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vid_v_u32m4(op0) \ +__builtin_rvv_vid_v_u32m4((size_t)(op0)) +#define vid_v_u32m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vid_v_u32m8(op0) \ +__builtin_rvv_vid_v_u32m8((size_t)(op0)) +#define vid_v_u32m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vid_v_u32mf2(op0) \ +__builtin_rvv_vid_v_u32mf2((size_t)(op0)) +#define vid_v_u32mf2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vid_v_u64m1(op0) \ +__builtin_rvv_vid_v_u64m1((size_t)(op0)) +#define vid_v_u64m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vid_v_u64m2(op0) \ +__builtin_rvv_vid_v_u64m2((size_t)(op0)) +#define vid_v_u64m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vid_v_u64m4(op0) \ +__builtin_rvv_vid_v_u64m4((size_t)(op0)) +#define vid_v_u64m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vid_v_u64m8(op0) \ +__builtin_rvv_vid_v_u64m8((size_t)(op0)) +#define vid_v_u64m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmv_x_s_i8m1_i8(op0) \ +__builtin_rvv_vmv_x_s_i8m1_i8((vint8m1_t)(op0)) +#define vmv_x_s_i8m2_i8(op0) \ +__builtin_rvv_vmv_x_s_i8m2_i8((vint8m2_t)(op0)) +#define vmv_x_s_i8m4_i8(op0) \ +__builtin_rvv_vmv_x_s_i8m4_i8((vint8m4_t)(op0)) +#define vmv_x_s_i8m8_i8(op0) \ +__builtin_rvv_vmv_x_s_i8m8_i8((vint8m8_t)(op0)) +#define vmv_x_s_i8mf2_i8(op0) \ +__builtin_rvv_vmv_x_s_i8mf2_i8((vint8mf2_t)(op0)) +#define vmv_x_s_i8mf4_i8(op0) \ +__builtin_rvv_vmv_x_s_i8mf4_i8((vint8mf4_t)(op0)) +#define vmv_x_s_i8mf8_i8(op0) \ +__builtin_rvv_vmv_x_s_i8mf8_i8((vint8mf8_t)(op0)) +#define vmv_x_s_i16m1_i16(op0) \ +__builtin_rvv_vmv_x_s_i16m1_i16((vint16m1_t)(op0)) +#define vmv_x_s_i16m2_i16(op0) \ +__builtin_rvv_vmv_x_s_i16m2_i16((vint16m2_t)(op0)) +#define vmv_x_s_i16m4_i16(op0) \ +__builtin_rvv_vmv_x_s_i16m4_i16((vint16m4_t)(op0)) +#define vmv_x_s_i16m8_i16(op0) \ +__builtin_rvv_vmv_x_s_i16m8_i16((vint16m8_t)(op0)) +#define vmv_x_s_i16mf2_i16(op0) \ +__builtin_rvv_vmv_x_s_i16mf2_i16((vint16mf2_t)(op0)) +#define vmv_x_s_i16mf4_i16(op0) \ +__builtin_rvv_vmv_x_s_i16mf4_i16((vint16mf4_t)(op0)) +#define vmv_x_s_i32m1_i32(op0) \ +__builtin_rvv_vmv_x_s_i32m1_i32((vint32m1_t)(op0)) +#define vmv_x_s_i32m2_i32(op0) \ +__builtin_rvv_vmv_x_s_i32m2_i32((vint32m2_t)(op0)) +#define vmv_x_s_i32m4_i32(op0) \ +__builtin_rvv_vmv_x_s_i32m4_i32((vint32m4_t)(op0)) +#define vmv_x_s_i32m8_i32(op0) \ +__builtin_rvv_vmv_x_s_i32m8_i32((vint32m8_t)(op0)) +#define vmv_x_s_i32mf2_i32(op0) \ +__builtin_rvv_vmv_x_s_i32mf2_i32((vint32mf2_t)(op0)) +#define vmv_x_s_i64m1_i64(op0) \ +__builtin_rvv_vmv_x_s_i64m1_i64((vint64m1_t)(op0)) +#define vmv_x_s_i64m2_i64(op0) \ +__builtin_rvv_vmv_x_s_i64m2_i64((vint64m2_t)(op0)) +#define vmv_x_s_i64m4_i64(op0) \ +__builtin_rvv_vmv_x_s_i64m4_i64((vint64m4_t)(op0)) +#define vmv_x_s_i64m8_i64(op0) \ +__builtin_rvv_vmv_x_s_i64m8_i64((vint64m8_t)(op0)) +#define vmv_x_s_u8m1_u8(op0) \ +__builtin_rvv_vmv_x_s_u8m1_u8((vuint8m1_t)(op0)) +#define vmv_x_s_u8m2_u8(op0) \ +__builtin_rvv_vmv_x_s_u8m2_u8((vuint8m2_t)(op0)) +#define vmv_x_s_u8m4_u8(op0) \ +__builtin_rvv_vmv_x_s_u8m4_u8((vuint8m4_t)(op0)) +#define vmv_x_s_u8m8_u8(op0) \ +__builtin_rvv_vmv_x_s_u8m8_u8((vuint8m8_t)(op0)) +#define vmv_x_s_u8mf2_u8(op0) \ +__builtin_rvv_vmv_x_s_u8mf2_u8((vuint8mf2_t)(op0)) +#define vmv_x_s_u8mf4_u8(op0) \ +__builtin_rvv_vmv_x_s_u8mf4_u8((vuint8mf4_t)(op0)) +#define vmv_x_s_u8mf8_u8(op0) \ +__builtin_rvv_vmv_x_s_u8mf8_u8((vuint8mf8_t)(op0)) +#define vmv_x_s_u16m1_u16(op0) \ +__builtin_rvv_vmv_x_s_u16m1_u16((vuint16m1_t)(op0)) +#define vmv_x_s_u16m2_u16(op0) \ +__builtin_rvv_vmv_x_s_u16m2_u16((vuint16m2_t)(op0)) +#define vmv_x_s_u16m4_u16(op0) \ +__builtin_rvv_vmv_x_s_u16m4_u16((vuint16m4_t)(op0)) +#define vmv_x_s_u16m8_u16(op0) \ +__builtin_rvv_vmv_x_s_u16m8_u16((vuint16m8_t)(op0)) +#define vmv_x_s_u16mf2_u16(op0) \ +__builtin_rvv_vmv_x_s_u16mf2_u16((vuint16mf2_t)(op0)) +#define vmv_x_s_u16mf4_u16(op0) \ +__builtin_rvv_vmv_x_s_u16mf4_u16((vuint16mf4_t)(op0)) +#define vmv_x_s_u32m1_u32(op0) \ +__builtin_rvv_vmv_x_s_u32m1_u32((vuint32m1_t)(op0)) +#define vmv_x_s_u32m2_u32(op0) \ +__builtin_rvv_vmv_x_s_u32m2_u32((vuint32m2_t)(op0)) +#define vmv_x_s_u32m4_u32(op0) \ +__builtin_rvv_vmv_x_s_u32m4_u32((vuint32m4_t)(op0)) +#define vmv_x_s_u32m8_u32(op0) \ +__builtin_rvv_vmv_x_s_u32m8_u32((vuint32m8_t)(op0)) +#define vmv_x_s_u32mf2_u32(op0) \ +__builtin_rvv_vmv_x_s_u32mf2_u32((vuint32mf2_t)(op0)) +#define vmv_x_s_u64m1_u64(op0) \ +__builtin_rvv_vmv_x_s_u64m1_u64((vuint64m1_t)(op0)) +#define vmv_x_s_u64m2_u64(op0) \ +__builtin_rvv_vmv_x_s_u64m2_u64((vuint64m2_t)(op0)) +#define vmv_x_s_u64m4_u64(op0) \ +__builtin_rvv_vmv_x_s_u64m4_u64((vuint64m4_t)(op0)) +#define vmv_x_s_u64m8_u64(op0) \ +__builtin_rvv_vmv_x_s_u64m8_u64((vuint64m8_t)(op0)) +#define vmv_s_x_i8m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i8m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i8m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i8m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i16m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmv_s_x_i16m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmv_s_x_i16m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmv_s_x_i16m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmv_s_x_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmv_s_x_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmv_s_x_i32m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmv_s_x_i32m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmv_s_x_i32m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmv_s_x_i32m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmv_s_x_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmv_s_x_i64m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmv_s_x_i64m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmv_s_x_i64m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmv_s_x_i64m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u16m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmv_s_x_u16m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmv_s_x_u16m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmv_s_x_u16m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmv_s_x_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmv_s_x_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmv_s_x_u32m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmv_s_x_u32m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmv_s_x_u32m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmv_s_x_u32m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmv_s_x_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmv_s_x_u64m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmv_s_x_u64m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmv_s_x_u64m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmv_s_x_u64m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u16m1(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u16m2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u16m4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u16m8(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u16mf2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u16mf4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u32m1(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u32m2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u32m4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u32m8(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u32mf2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u64m1(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u64m2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u64m4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u64m8(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vslideup_vx_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8m1((vuint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8m2((vuint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8m4((vuint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8m8((vuint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8mf2((vuint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8mf4((vuint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8mf8((vuint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u16m1((vuint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u16m2((vuint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u16m4((vuint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u16m8((vuint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u16mf2((vuint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u16mf4((vuint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u32m1((vuint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u32m2((vuint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u32m4((vuint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u32m8((vuint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u32mf2((vuint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u64m1((vuint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u64m2((vuint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u64m4((vuint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u64m8((vuint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8m1((vuint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8m2((vuint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8m4((vuint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8m8((vuint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8mf2((vuint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8mf4((vuint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8mf8((vuint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u16m1((vuint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u16m2((vuint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u16m4((vuint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u16m8((vuint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u16mf2((vuint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u16mf4((vuint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u32m1((vuint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u32m2((vuint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u32m4((vuint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u32m8((vuint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u32mf2((vuint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u64m1((vuint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u64m2((vuint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u64m4((vuint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u64m8((vuint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8m1((vint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8m2((vint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8m4((vint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8m8((vint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8mf2((vint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8mf4((vint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8mf8((vint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vrgather_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i16m1((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i16m2((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i16m4((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i16m8((vint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i16mf2((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i16mf4((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgather_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgather_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i32m1((vint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i32m2((vint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i32m4((vint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i32m8((vint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i32mf2((vint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i64m1((vint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i64m2((vint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i64m4((vint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i64m8((vint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8m1((vint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8m2((vint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8m4((vint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8m8((vint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8mf2((vint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8mf4((vint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8mf8((vint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i16m1((vint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i16m2((vint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i16m4((vint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i16m8((vint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i16mf2((vint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i16mf4((vint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i32m1((vint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i32m2((vint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i32m4((vint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i32m8((vint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i32mf2((vint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i64m1((vint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i64m2((vint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i64m4((vint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i64m8((vint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i8m1((vint8m1_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i8m2((vint8m2_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i8m4((vint8m4_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i8mf2((vint8mf2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i8mf4((vint8mf4_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i8mf8((vint8mf8_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i16m1((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i16m2((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i16m4((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i16m8((vint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i16mf2((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i16mf4((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i32m1((vint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i32m2((vint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i32m4((vint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i32m8((vint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i32mf2((vint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i64m1((vint64m1_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i64m2((vint64m2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i64m4((vint64m4_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i64m8((vint64m8_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vrgather_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgather_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgather_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8m1((vuint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8m2((vuint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8m4((vuint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8m8((vuint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8mf2((vuint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8mf4((vuint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8mf8((vuint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u16m1((vuint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u16m2((vuint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u16m4((vuint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u16m8((vuint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u16mf2((vuint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u16mf4((vuint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u32m1((vuint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u32m2((vuint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u32m4((vuint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u32m8((vuint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u32mf2((vuint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u64m1((vuint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u64m2((vuint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u64m4((vuint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u64m8((vuint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u8m1((vuint8m1_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u8m2((vuint8m2_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u8m4((vuint8m4_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u8mf2((vuint8mf2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u8mf4((vuint8mf4_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u8mf8((vuint8mf8_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u32m1((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u32m2((vuint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u32m4((vuint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u32m8((vuint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u32mf2((vuint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u64m1((vuint64m1_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u64m2((vuint64m2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u64m4((vuint64m4_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u64m8((vuint64m8_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vcompress_vm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8m1((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8m2((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8m4((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8m8((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8mf2((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8mf4((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vcompress_vm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8mf8((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vcompress_vm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i16m1((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i16m2((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i16m4((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i16m8((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i16mf2((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i16mf4((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vcompress_vm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i32m1((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i32m2((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i32m4((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i32m8((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i32mf2((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i64m1((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i64m2((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i64m4((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i64m8((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8m1((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8m2((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8m4((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8m8((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8mf2((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8mf4((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8mf8((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vcompress_vm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u16m1((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u16m2((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u16m4((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u16m8((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u16mf2((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u16mf4((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vcompress_vm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u32m1((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u32m2((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u32m4((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u32m8((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u32mf2((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u64m1((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u64m2((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u64m4((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u64m8((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vget_v_i8m2_i8m1(op0, op1) \ +__builtin_rvv_vget_v_i8m2_i8m1((vint8m2_t)(op0), (size_t)(op1)) +#define vget_v_i8m4_i8m1(op0, op1) \ +__builtin_rvv_vget_v_i8m4_i8m1((vint8m4_t)(op0), (size_t)(op1)) +#define vget_v_i8m8_i8m1(op0, op1) \ +__builtin_rvv_vget_v_i8m8_i8m1((vint8m8_t)(op0), (size_t)(op1)) +#define vget_v_i16m2_i16m1(op0, op1) \ +__builtin_rvv_vget_v_i16m2_i16m1((vint16m2_t)(op0), (size_t)(op1)) +#define vget_v_i16m4_i16m1(op0, op1) \ +__builtin_rvv_vget_v_i16m4_i16m1((vint16m4_t)(op0), (size_t)(op1)) +#define vget_v_i16m8_i16m1(op0, op1) \ +__builtin_rvv_vget_v_i16m8_i16m1((vint16m8_t)(op0), (size_t)(op1)) +#define vget_v_i32m2_i32m1(op0, op1) \ +__builtin_rvv_vget_v_i32m2_i32m1((vint32m2_t)(op0), (size_t)(op1)) +#define vget_v_i32m4_i32m1(op0, op1) \ +__builtin_rvv_vget_v_i32m4_i32m1((vint32m4_t)(op0), (size_t)(op1)) +#define vget_v_i32m8_i32m1(op0, op1) \ +__builtin_rvv_vget_v_i32m8_i32m1((vint32m8_t)(op0), (size_t)(op1)) +#define vget_v_i64m2_i64m1(op0, op1) \ +__builtin_rvv_vget_v_i64m2_i64m1((vint64m2_t)(op0), (size_t)(op1)) +#define vget_v_i64m4_i64m1(op0, op1) \ +__builtin_rvv_vget_v_i64m4_i64m1((vint64m4_t)(op0), (size_t)(op1)) +#define vget_v_i64m8_i64m1(op0, op1) \ +__builtin_rvv_vget_v_i64m8_i64m1((vint64m8_t)(op0), (size_t)(op1)) +#define vget_v_u8m2_u8m1(op0, op1) \ +__builtin_rvv_vget_v_u8m2_u8m1((vuint8m2_t)(op0), (size_t)(op1)) +#define vget_v_u8m4_u8m1(op0, op1) \ +__builtin_rvv_vget_v_u8m4_u8m1((vuint8m4_t)(op0), (size_t)(op1)) +#define vget_v_u8m8_u8m1(op0, op1) \ +__builtin_rvv_vget_v_u8m8_u8m1((vuint8m8_t)(op0), (size_t)(op1)) +#define vget_v_u16m2_u16m1(op0, op1) \ +__builtin_rvv_vget_v_u16m2_u16m1((vuint16m2_t)(op0), (size_t)(op1)) +#define vget_v_u16m4_u16m1(op0, op1) \ +__builtin_rvv_vget_v_u16m4_u16m1((vuint16m4_t)(op0), (size_t)(op1)) +#define vget_v_u16m8_u16m1(op0, op1) \ +__builtin_rvv_vget_v_u16m8_u16m1((vuint16m8_t)(op0), (size_t)(op1)) +#define vget_v_u32m2_u32m1(op0, op1) \ +__builtin_rvv_vget_v_u32m2_u32m1((vuint32m2_t)(op0), (size_t)(op1)) +#define vget_v_u32m4_u32m1(op0, op1) \ +__builtin_rvv_vget_v_u32m4_u32m1((vuint32m4_t)(op0), (size_t)(op1)) +#define vget_v_u32m8_u32m1(op0, op1) \ +__builtin_rvv_vget_v_u32m8_u32m1((vuint32m8_t)(op0), (size_t)(op1)) +#define vget_v_u64m2_u64m1(op0, op1) \ +__builtin_rvv_vget_v_u64m2_u64m1((vuint64m2_t)(op0), (size_t)(op1)) +#define vget_v_u64m4_u64m1(op0, op1) \ +__builtin_rvv_vget_v_u64m4_u64m1((vuint64m4_t)(op0), (size_t)(op1)) +#define vget_v_u64m8_u64m1(op0, op1) \ +__builtin_rvv_vget_v_u64m8_u64m1((vuint64m8_t)(op0), (size_t)(op1)) +#define vget_v_i8m4_i8m2(op0, op1) \ +__builtin_rvv_vget_v_i8m4_i8m2((vint8m4_t)(op0), (size_t)(op1)) +#define vget_v_i8m8_i8m2(op0, op1) \ +__builtin_rvv_vget_v_i8m8_i8m2((vint8m8_t)(op0), (size_t)(op1)) +#define vget_v_i16m4_i16m2(op0, op1) \ +__builtin_rvv_vget_v_i16m4_i16m2((vint16m4_t)(op0), (size_t)(op1)) +#define vget_v_i16m8_i16m2(op0, op1) \ +__builtin_rvv_vget_v_i16m8_i16m2((vint16m8_t)(op0), (size_t)(op1)) +#define vget_v_i32m4_i32m2(op0, op1) \ +__builtin_rvv_vget_v_i32m4_i32m2((vint32m4_t)(op0), (size_t)(op1)) +#define vget_v_i32m8_i32m2(op0, op1) \ +__builtin_rvv_vget_v_i32m8_i32m2((vint32m8_t)(op0), (size_t)(op1)) +#define vget_v_i64m4_i64m2(op0, op1) \ +__builtin_rvv_vget_v_i64m4_i64m2((vint64m4_t)(op0), (size_t)(op1)) +#define vget_v_i64m8_i64m2(op0, op1) \ +__builtin_rvv_vget_v_i64m8_i64m2((vint64m8_t)(op0), (size_t)(op1)) +#define vget_v_u8m4_u8m2(op0, op1) \ +__builtin_rvv_vget_v_u8m4_u8m2((vuint8m4_t)(op0), (size_t)(op1)) +#define vget_v_u8m8_u8m2(op0, op1) \ +__builtin_rvv_vget_v_u8m8_u8m2((vuint8m8_t)(op0), (size_t)(op1)) +#define vget_v_u16m4_u16m2(op0, op1) \ +__builtin_rvv_vget_v_u16m4_u16m2((vuint16m4_t)(op0), (size_t)(op1)) +#define vget_v_u16m8_u16m2(op0, op1) \ +__builtin_rvv_vget_v_u16m8_u16m2((vuint16m8_t)(op0), (size_t)(op1)) +#define vget_v_u32m4_u32m2(op0, op1) \ +__builtin_rvv_vget_v_u32m4_u32m2((vuint32m4_t)(op0), (size_t)(op1)) +#define vget_v_u32m8_u32m2(op0, op1) \ +__builtin_rvv_vget_v_u32m8_u32m2((vuint32m8_t)(op0), (size_t)(op1)) +#define vget_v_u64m4_u64m2(op0, op1) \ +__builtin_rvv_vget_v_u64m4_u64m2((vuint64m4_t)(op0), (size_t)(op1)) +#define vget_v_u64m8_u64m2(op0, op1) \ +__builtin_rvv_vget_v_u64m8_u64m2((vuint64m8_t)(op0), (size_t)(op1)) +#define vget_v_i8m8_i8m4(op0, op1) \ +__builtin_rvv_vget_v_i8m8_i8m4((vint8m8_t)(op0), (size_t)(op1)) +#define vget_v_i16m8_i16m4(op0, op1) \ +__builtin_rvv_vget_v_i16m8_i16m4((vint16m8_t)(op0), (size_t)(op1)) +#define vget_v_i32m8_i32m4(op0, op1) \ +__builtin_rvv_vget_v_i32m8_i32m4((vint32m8_t)(op0), (size_t)(op1)) +#define vget_v_i64m8_i64m4(op0, op1) \ +__builtin_rvv_vget_v_i64m8_i64m4((vint64m8_t)(op0), (size_t)(op1)) +#define vget_v_u8m8_u8m4(op0, op1) \ +__builtin_rvv_vget_v_u8m8_u8m4((vuint8m8_t)(op0), (size_t)(op1)) +#define vget_v_u16m8_u16m4(op0, op1) \ +__builtin_rvv_vget_v_u16m8_u16m4((vuint16m8_t)(op0), (size_t)(op1)) +#define vget_v_u32m8_u32m4(op0, op1) \ +__builtin_rvv_vget_v_u32m8_u32m4((vuint32m8_t)(op0), (size_t)(op1)) +#define vget_v_u64m8_u64m4(op0, op1) \ +__builtin_rvv_vget_v_u64m8_u64m4((vuint64m8_t)(op0), (size_t)(op1)) +#define vset_v_i8m1_i8m2(op0, op1, op2) \ +__builtin_rvv_vset_v_i8m1_i8m2((vint8m2_t)(op0), (size_t)(op1), (vint8m1_t)(op2)) +#define vset_v_i16m1_i16m2(op0, op1, op2) \ +__builtin_rvv_vset_v_i16m1_i16m2((vint16m2_t)(op0), (size_t)(op1), (vint16m1_t)(op2)) +#define vset_v_i32m1_i32m2(op0, op1, op2) \ +__builtin_rvv_vset_v_i32m1_i32m2((vint32m2_t)(op0), (size_t)(op1), (vint32m1_t)(op2)) +#define vset_v_i64m1_i64m2(op0, op1, op2) \ +__builtin_rvv_vset_v_i64m1_i64m2((vint64m2_t)(op0), (size_t)(op1), (vint64m1_t)(op2)) +#define vset_v_u8m1_u8m2(op0, op1, op2) \ +__builtin_rvv_vset_v_u8m1_u8m2((vuint8m2_t)(op0), (size_t)(op1), (vuint8m1_t)(op2)) +#define vset_v_u16m1_u16m2(op0, op1, op2) \ +__builtin_rvv_vset_v_u16m1_u16m2((vuint16m2_t)(op0), (size_t)(op1), (vuint16m1_t)(op2)) +#define vset_v_u32m1_u32m2(op0, op1, op2) \ +__builtin_rvv_vset_v_u32m1_u32m2((vuint32m2_t)(op0), (size_t)(op1), (vuint32m1_t)(op2)) +#define vset_v_u64m1_u64m2(op0, op1, op2) \ +__builtin_rvv_vset_v_u64m1_u64m2((vuint64m2_t)(op0), (size_t)(op1), (vuint64m1_t)(op2)) +#define vset_v_i8m1_i8m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i8m1_i8m4((vint8m4_t)(op0), (size_t)(op1), (vint8m1_t)(op2)) +#define vset_v_i8m2_i8m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i8m2_i8m4((vint8m4_t)(op0), (size_t)(op1), (vint8m2_t)(op2)) +#define vset_v_i16m1_i16m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i16m1_i16m4((vint16m4_t)(op0), (size_t)(op1), (vint16m1_t)(op2)) +#define vset_v_i16m2_i16m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i16m2_i16m4((vint16m4_t)(op0), (size_t)(op1), (vint16m2_t)(op2)) +#define vset_v_i32m1_i32m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i32m1_i32m4((vint32m4_t)(op0), (size_t)(op1), (vint32m1_t)(op2)) +#define vset_v_i32m2_i32m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i32m2_i32m4((vint32m4_t)(op0), (size_t)(op1), (vint32m2_t)(op2)) +#define vset_v_i64m1_i64m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i64m1_i64m4((vint64m4_t)(op0), (size_t)(op1), (vint64m1_t)(op2)) +#define vset_v_i64m2_i64m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i64m2_i64m4((vint64m4_t)(op0), (size_t)(op1), (vint64m2_t)(op2)) +#define vset_v_u8m1_u8m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u8m1_u8m4((vuint8m4_t)(op0), (size_t)(op1), (vuint8m1_t)(op2)) +#define vset_v_u8m2_u8m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u8m2_u8m4((vuint8m4_t)(op0), (size_t)(op1), (vuint8m2_t)(op2)) +#define vset_v_u16m1_u16m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u16m1_u16m4((vuint16m4_t)(op0), (size_t)(op1), (vuint16m1_t)(op2)) +#define vset_v_u16m2_u16m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u16m2_u16m4((vuint16m4_t)(op0), (size_t)(op1), (vuint16m2_t)(op2)) +#define vset_v_u32m1_u32m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u32m1_u32m4((vuint32m4_t)(op0), (size_t)(op1), (vuint32m1_t)(op2)) +#define vset_v_u32m2_u32m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u32m2_u32m4((vuint32m4_t)(op0), (size_t)(op1), (vuint32m2_t)(op2)) +#define vset_v_u64m1_u64m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u64m1_u64m4((vuint64m4_t)(op0), (size_t)(op1), (vuint64m1_t)(op2)) +#define vset_v_u64m2_u64m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u64m2_u64m4((vuint64m4_t)(op0), (size_t)(op1), (vuint64m2_t)(op2)) +#define vset_v_i8m1_i8m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i8m1_i8m8((vint8m8_t)(op0), (size_t)(op1), (vint8m1_t)(op2)) +#define vset_v_i8m2_i8m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i8m2_i8m8((vint8m8_t)(op0), (size_t)(op1), (vint8m2_t)(op2)) +#define vset_v_i8m4_i8m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i8m4_i8m8((vint8m8_t)(op0), (size_t)(op1), (vint8m4_t)(op2)) +#define vset_v_i16m1_i16m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i16m1_i16m8((vint16m8_t)(op0), (size_t)(op1), (vint16m1_t)(op2)) +#define vset_v_i16m2_i16m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i16m2_i16m8((vint16m8_t)(op0), (size_t)(op1), (vint16m2_t)(op2)) +#define vset_v_i16m4_i16m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i16m4_i16m8((vint16m8_t)(op0), (size_t)(op1), (vint16m4_t)(op2)) +#define vset_v_i32m1_i32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i32m1_i32m8((vint32m8_t)(op0), (size_t)(op1), (vint32m1_t)(op2)) +#define vset_v_i32m2_i32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i32m2_i32m8((vint32m8_t)(op0), (size_t)(op1), (vint32m2_t)(op2)) +#define vset_v_i32m4_i32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i32m4_i32m8((vint32m8_t)(op0), (size_t)(op1), (vint32m4_t)(op2)) +#define vset_v_i64m1_i64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i64m1_i64m8((vint64m8_t)(op0), (size_t)(op1), (vint64m1_t)(op2)) +#define vset_v_i64m2_i64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i64m2_i64m8((vint64m8_t)(op0), (size_t)(op1), (vint64m2_t)(op2)) +#define vset_v_i64m4_i64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i64m4_i64m8((vint64m8_t)(op0), (size_t)(op1), (vint64m4_t)(op2)) +#define vset_v_u8m1_u8m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u8m1_u8m8((vuint8m8_t)(op0), (size_t)(op1), (vuint8m1_t)(op2)) +#define vset_v_u8m2_u8m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u8m2_u8m8((vuint8m8_t)(op0), (size_t)(op1), (vuint8m2_t)(op2)) +#define vset_v_u8m4_u8m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u8m4_u8m8((vuint8m8_t)(op0), (size_t)(op1), (vuint8m4_t)(op2)) +#define vset_v_u16m1_u16m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u16m1_u16m8((vuint16m8_t)(op0), (size_t)(op1), (vuint16m1_t)(op2)) +#define vset_v_u16m2_u16m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u16m2_u16m8((vuint16m8_t)(op0), (size_t)(op1), (vuint16m2_t)(op2)) +#define vset_v_u16m4_u16m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u16m4_u16m8((vuint16m8_t)(op0), (size_t)(op1), (vuint16m4_t)(op2)) +#define vset_v_u32m1_u32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u32m1_u32m8((vuint32m8_t)(op0), (size_t)(op1), (vuint32m1_t)(op2)) +#define vset_v_u32m2_u32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u32m2_u32m8((vuint32m8_t)(op0), (size_t)(op1), (vuint32m2_t)(op2)) +#define vset_v_u32m4_u32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u32m4_u32m8((vuint32m8_t)(op0), (size_t)(op1), (vuint32m4_t)(op2)) +#define vset_v_u64m1_u64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u64m1_u64m8((vuint64m8_t)(op0), (size_t)(op1), (vuint64m1_t)(op2)) +#define vset_v_u64m2_u64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u64m2_u64m8((vuint64m8_t)(op0), (size_t)(op1), (vuint64m2_t)(op2)) +#define vset_v_u64m4_u64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u64m4_u64m8((vuint64m8_t)(op0), (size_t)(op1), (vuint64m4_t)(op2)) +#define vsoxei8_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8m4((int8_t *)(op0), (vuint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8m8((int8_t *)(op0), (vuint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8m8_m((vbool1_t)(op0), (int8_t *)(op1), (vuint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8m4((uint8_t *)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8m8((uint8_t *)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8m8_m((vbool1_t)(op0), (uint8_t *)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i8m4((int8_t *)(op0), (vuint16m8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint16m8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u8m4((uint8_t *)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i16m4((int16_t *)(op0), (vuint8m2_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint8m2_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i16m8((int16_t *)(op0), (vuint8m4_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i16m8_m((vbool2_t)(op0), (int16_t *)(op1), (vuint8m4_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u16m4((uint16_t *)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint8m2_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u16m8((uint16_t *)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u16m8_m((vbool2_t)(op0), (uint16_t *)(op1), (vuint8m4_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i16m4((int16_t *)(op0), (vuint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i16m8((int16_t *)(op0), (vuint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i16m8_m((vbool2_t)(op0), (int16_t *)(op1), (vuint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u16m4((uint16_t *)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u16m8((uint16_t *)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u16m8_m((vbool2_t)(op0), (uint16_t *)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i16m4((int16_t *)(op0), (vuint32m8_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint32m8_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u16m4((uint16_t *)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vle16ff_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_i16m1((const int16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_i16m2((const int16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_i16m4((const int16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_i16m8((const int16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_i16mf2((const int16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_i16mf4((const int16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_u16m1((const uint16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_u16m2((const uint16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_u16m4((const uint16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_u16m8((const uint16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_u16mf2((const uint16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_u16mf4((const uint16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_i32m1((const int32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_i32m2((const int32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_i32m4((const int32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_i32m8((const int32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_i32mf2((const int32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_u32m1((const uint32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_u32m2((const uint32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_u32m4((const uint32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_u32m8((const uint32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_u32mf2((const uint32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_i64m1((const int64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_i64m2((const int64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_i64m4((const int64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_i64m8((const int64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_u64m1((const uint64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_u64m2((const uint64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_u64m4((const uint64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_u64m8((const uint64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8_v_i8m1(op0, op1) \ +__builtin_rvv_vle8_v_i8m1((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_i8m2(op0, op1) \ +__builtin_rvv_vle8_v_i8m2((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_i8m4(op0, op1) \ +__builtin_rvv_vle8_v_i8m4((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_i8m8(op0, op1) \ +__builtin_rvv_vle8_v_i8m8((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_i8mf2(op0, op1) \ +__builtin_rvv_vle8_v_i8mf2((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_i8mf4(op0, op1) \ +__builtin_rvv_vle8_v_i8mf4((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_i8mf8(op0, op1) \ +__builtin_rvv_vle8_v_i8mf8((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8m1(op0, op1) \ +__builtin_rvv_vle8_v_u8m1((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8m2(op0, op1) \ +__builtin_rvv_vle8_v_u8m2((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8m4(op0, op1) \ +__builtin_rvv_vle8_v_u8m4((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8m8(op0, op1) \ +__builtin_rvv_vle8_v_u8m8((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8mf2(op0, op1) \ +__builtin_rvv_vle8_v_u8mf2((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8mf4(op0, op1) \ +__builtin_rvv_vle8_v_u8mf4((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8mf8(op0, op1) \ +__builtin_rvv_vle8_v_u8mf8((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8ff_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8m1((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8m2((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8m4((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_i8m8(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8m8((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8mf2((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8mf4((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8mf8((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8m1((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8m2((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8m4((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8m8(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8m8((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8mf2((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8mf4((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8mf8((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vneg_v_i8m1(op0, op1) \ +__builtin_rvv_vneg_v_i8m1((vint8m1_t)(op0), (size_t)(op1)) +#define vneg_v_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vneg_v_i8m2(op0, op1) \ +__builtin_rvv_vneg_v_i8m2((vint8m2_t)(op0), (size_t)(op1)) +#define vneg_v_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vneg_v_i8m4(op0, op1) \ +__builtin_rvv_vneg_v_i8m4((vint8m4_t)(op0), (size_t)(op1)) +#define vneg_v_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vneg_v_i8m8(op0, op1) \ +__builtin_rvv_vneg_v_i8m8((vint8m8_t)(op0), (size_t)(op1)) +#define vneg_v_i8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vneg_v_i8mf2(op0, op1) \ +__builtin_rvv_vneg_v_i8mf2((vint8mf2_t)(op0), (size_t)(op1)) +#define vneg_v_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vneg_v_i8mf4(op0, op1) \ +__builtin_rvv_vneg_v_i8mf4((vint8mf4_t)(op0), (size_t)(op1)) +#define vneg_v_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vneg_v_i8mf8(op0, op1) \ +__builtin_rvv_vneg_v_i8mf8((vint8mf8_t)(op0), (size_t)(op1)) +#define vneg_v_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vneg_v_i16m1(op0, op1) \ +__builtin_rvv_vneg_v_i16m1((vint16m1_t)(op0), (size_t)(op1)) +#define vneg_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vneg_v_i16m2(op0, op1) \ +__builtin_rvv_vneg_v_i16m2((vint16m2_t)(op0), (size_t)(op1)) +#define vneg_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vneg_v_i16m4(op0, op1) \ +__builtin_rvv_vneg_v_i16m4((vint16m4_t)(op0), (size_t)(op1)) +#define vneg_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vneg_v_i16m8(op0, op1) \ +__builtin_rvv_vneg_v_i16m8((vint16m8_t)(op0), (size_t)(op1)) +#define vneg_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vneg_v_i16mf2(op0, op1) \ +__builtin_rvv_vneg_v_i16mf2((vint16mf2_t)(op0), (size_t)(op1)) +#define vneg_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vneg_v_i16mf4(op0, op1) \ +__builtin_rvv_vneg_v_i16mf4((vint16mf4_t)(op0), (size_t)(op1)) +#define vneg_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vneg_v_i32m1(op0, op1) \ +__builtin_rvv_vneg_v_i32m1((vint32m1_t)(op0), (size_t)(op1)) +#define vneg_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vneg_v_i32m2(op0, op1) \ +__builtin_rvv_vneg_v_i32m2((vint32m2_t)(op0), (size_t)(op1)) +#define vneg_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vneg_v_i32m4(op0, op1) \ +__builtin_rvv_vneg_v_i32m4((vint32m4_t)(op0), (size_t)(op1)) +#define vneg_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vneg_v_i32m8(op0, op1) \ +__builtin_rvv_vneg_v_i32m8((vint32m8_t)(op0), (size_t)(op1)) +#define vneg_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vneg_v_i32mf2(op0, op1) \ +__builtin_rvv_vneg_v_i32mf2((vint32mf2_t)(op0), (size_t)(op1)) +#define vneg_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vneg_v_i64m1(op0, op1) \ +__builtin_rvv_vneg_v_i64m1((vint64m1_t)(op0), (size_t)(op1)) +#define vneg_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vneg_v_i64m2(op0, op1) \ +__builtin_rvv_vneg_v_i64m2((vint64m2_t)(op0), (size_t)(op1)) +#define vneg_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vneg_v_i64m4(op0, op1) \ +__builtin_rvv_vneg_v_i64m4((vint64m4_t)(op0), (size_t)(op1)) +#define vneg_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vneg_v_i64m8(op0, op1) \ +__builtin_rvv_vneg_v_i64m8((vint64m8_t)(op0), (size_t)(op1)) +#define vneg_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vnot_v_i8m1(op0, op1) \ +__builtin_rvv_vnot_v_i8m1((vint8m1_t)(op0), (size_t)(op1)) +#define vnot_v_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vnot_v_i8m2(op0, op1) \ +__builtin_rvv_vnot_v_i8m2((vint8m2_t)(op0), (size_t)(op1)) +#define vnot_v_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vnot_v_i8m4(op0, op1) \ +__builtin_rvv_vnot_v_i8m4((vint8m4_t)(op0), (size_t)(op1)) +#define vnot_v_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vnot_v_i8m8(op0, op1) \ +__builtin_rvv_vnot_v_i8m8((vint8m8_t)(op0), (size_t)(op1)) +#define vnot_v_i8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vnot_v_i8mf2(op0, op1) \ +__builtin_rvv_vnot_v_i8mf2((vint8mf2_t)(op0), (size_t)(op1)) +#define vnot_v_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vnot_v_i8mf4(op0, op1) \ +__builtin_rvv_vnot_v_i8mf4((vint8mf4_t)(op0), (size_t)(op1)) +#define vnot_v_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vnot_v_i8mf8(op0, op1) \ +__builtin_rvv_vnot_v_i8mf8((vint8mf8_t)(op0), (size_t)(op1)) +#define vnot_v_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vnot_v_i16m1(op0, op1) \ +__builtin_rvv_vnot_v_i16m1((vint16m1_t)(op0), (size_t)(op1)) +#define vnot_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vnot_v_i16m2(op0, op1) \ +__builtin_rvv_vnot_v_i16m2((vint16m2_t)(op0), (size_t)(op1)) +#define vnot_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vnot_v_i16m4(op0, op1) \ +__builtin_rvv_vnot_v_i16m4((vint16m4_t)(op0), (size_t)(op1)) +#define vnot_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vnot_v_i16m8(op0, op1) \ +__builtin_rvv_vnot_v_i16m8((vint16m8_t)(op0), (size_t)(op1)) +#define vnot_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vnot_v_i16mf2(op0, op1) \ +__builtin_rvv_vnot_v_i16mf2((vint16mf2_t)(op0), (size_t)(op1)) +#define vnot_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vnot_v_i16mf4(op0, op1) \ +__builtin_rvv_vnot_v_i16mf4((vint16mf4_t)(op0), (size_t)(op1)) +#define vnot_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vnot_v_i32m1(op0, op1) \ +__builtin_rvv_vnot_v_i32m1((vint32m1_t)(op0), (size_t)(op1)) +#define vnot_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vnot_v_i32m2(op0, op1) \ +__builtin_rvv_vnot_v_i32m2((vint32m2_t)(op0), (size_t)(op1)) +#define vnot_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vnot_v_i32m4(op0, op1) \ +__builtin_rvv_vnot_v_i32m4((vint32m4_t)(op0), (size_t)(op1)) +#define vnot_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vnot_v_i32m8(op0, op1) \ +__builtin_rvv_vnot_v_i32m8((vint32m8_t)(op0), (size_t)(op1)) +#define vnot_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vnot_v_i32mf2(op0, op1) \ +__builtin_rvv_vnot_v_i32mf2((vint32mf2_t)(op0), (size_t)(op1)) +#define vnot_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vnot_v_i64m1(op0, op1) \ +__builtin_rvv_vnot_v_i64m1((vint64m1_t)(op0), (size_t)(op1)) +#define vnot_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vnot_v_i64m2(op0, op1) \ +__builtin_rvv_vnot_v_i64m2((vint64m2_t)(op0), (size_t)(op1)) +#define vnot_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vnot_v_i64m4(op0, op1) \ +__builtin_rvv_vnot_v_i64m4((vint64m4_t)(op0), (size_t)(op1)) +#define vnot_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vnot_v_i64m8(op0, op1) \ +__builtin_rvv_vnot_v_i64m8((vint64m8_t)(op0), (size_t)(op1)) +#define vnot_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vlse8_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8m1((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8m2((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8m4((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_i8m8(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8m8((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8mf2((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8mf4((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8mf8((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vnot_v_u8m1(op0, op1) \ +__builtin_rvv_vnot_v_u8m1((vuint8m1_t)(op0), (size_t)(op1)) +#define vnot_v_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vnot_v_u8m2(op0, op1) \ +__builtin_rvv_vnot_v_u8m2((vuint8m2_t)(op0), (size_t)(op1)) +#define vnot_v_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vnot_v_u8m4(op0, op1) \ +__builtin_rvv_vnot_v_u8m4((vuint8m4_t)(op0), (size_t)(op1)) +#define vnot_v_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vnot_v_u8m8(op0, op1) \ +__builtin_rvv_vnot_v_u8m8((vuint8m8_t)(op0), (size_t)(op1)) +#define vnot_v_u8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vnot_v_u8mf2(op0, op1) \ +__builtin_rvv_vnot_v_u8mf2((vuint8mf2_t)(op0), (size_t)(op1)) +#define vnot_v_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vnot_v_u8mf4(op0, op1) \ +__builtin_rvv_vnot_v_u8mf4((vuint8mf4_t)(op0), (size_t)(op1)) +#define vnot_v_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vnot_v_u8mf8(op0, op1) \ +__builtin_rvv_vnot_v_u8mf8((vuint8mf8_t)(op0), (size_t)(op1)) +#define vnot_v_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vnot_v_u16m1(op0, op1) \ +__builtin_rvv_vnot_v_u16m1((vuint16m1_t)(op0), (size_t)(op1)) +#define vnot_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vnot_v_u16m2(op0, op1) \ +__builtin_rvv_vnot_v_u16m2((vuint16m2_t)(op0), (size_t)(op1)) +#define vnot_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vnot_v_u16m4(op0, op1) \ +__builtin_rvv_vnot_v_u16m4((vuint16m4_t)(op0), (size_t)(op1)) +#define vnot_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vnot_v_u16m8(op0, op1) \ +__builtin_rvv_vnot_v_u16m8((vuint16m8_t)(op0), (size_t)(op1)) +#define vnot_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vnot_v_u16mf2(op0, op1) \ +__builtin_rvv_vnot_v_u16mf2((vuint16mf2_t)(op0), (size_t)(op1)) +#define vnot_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vnot_v_u16mf4(op0, op1) \ +__builtin_rvv_vnot_v_u16mf4((vuint16mf4_t)(op0), (size_t)(op1)) +#define vnot_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vnot_v_u32m1(op0, op1) \ +__builtin_rvv_vnot_v_u32m1((vuint32m1_t)(op0), (size_t)(op1)) +#define vnot_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vnot_v_u32m2(op0, op1) \ +__builtin_rvv_vnot_v_u32m2((vuint32m2_t)(op0), (size_t)(op1)) +#define vnot_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vnot_v_u32m4(op0, op1) \ +__builtin_rvv_vnot_v_u32m4((vuint32m4_t)(op0), (size_t)(op1)) +#define vnot_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vnot_v_u32m8(op0, op1) \ +__builtin_rvv_vnot_v_u32m8((vuint32m8_t)(op0), (size_t)(op1)) +#define vnot_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vnot_v_u32mf2(op0, op1) \ +__builtin_rvv_vnot_v_u32mf2((vuint32mf2_t)(op0), (size_t)(op1)) +#define vnot_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vnot_v_u64m1(op0, op1) \ +__builtin_rvv_vnot_v_u64m1((vuint64m1_t)(op0), (size_t)(op1)) +#define vnot_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vnot_v_u64m2(op0, op1) \ +__builtin_rvv_vnot_v_u64m2((vuint64m2_t)(op0), (size_t)(op1)) +#define vnot_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vnot_v_u64m4(op0, op1) \ +__builtin_rvv_vnot_v_u64m4((vuint64m4_t)(op0), (size_t)(op1)) +#define vnot_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vnot_v_u64m8(op0, op1) \ +__builtin_rvv_vnot_v_u64m8((vuint64m8_t)(op0), (size_t)(op1)) +#define vnot_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vmmv_m_b8(op0, op1) \ +__builtin_rvv_vmmv_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vmmv_m_b4(op0, op1) \ +__builtin_rvv_vmmv_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vmmv_m_b2(op0, op1) \ +__builtin_rvv_vmmv_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vmmv_m_b1(op0, op1) \ +__builtin_rvv_vmmv_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vmmv_m_b16(op0, op1) \ +__builtin_rvv_vmmv_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vmmv_m_b32(op0, op1) \ +__builtin_rvv_vmmv_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vmmv_m_b64(op0, op1) \ +__builtin_rvv_vmmv_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16m2(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u16m2((vuint8m1_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u16m4(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u16m4((vuint8m2_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u16m8(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u16m8((vuint8m4_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u16m1(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u16m1((vuint8mf2_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u16mf2(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u16mf2((vuint8mf4_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u16mf4(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u16mf4((vuint8mf8_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u32m2(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u32m2((vuint16m1_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u32m4(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u32m4((vuint16m2_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u32m8(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u32m8((vuint16m4_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u32m1(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u32m1((vuint16mf2_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u32mf2(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u32mf2((vuint16mf4_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u64m2(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u64m2((vuint32m1_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u64m4(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u64m4((vuint32m2_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u64m8(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u64m8((vuint32m4_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u64m1(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u64m1((vuint32mf2_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i8m1(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i8m1((vint16m2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i8m2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i8m2((vint16m4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i8m4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i8m4((vint16m8_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i8mf2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i8mf2((vint16m1_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i8mf4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i8mf4((vint16mf2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i8mf8(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i8mf8((vint16mf4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i16m1(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i16m1((vint32m2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i16m2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i16m2((vint32m4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i16m4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i16m4((vint32m8_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i16mf2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i16mf2((vint32m1_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i16mf4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i16mf4((vint32mf2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i32m1(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i32m1((vint64m2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i32m2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i32m2((vint64m4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i32m4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i32m4((vint64m8_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i32mf2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i32mf2((vint64m1_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vle16_v_i16m1(op0, op1) \ +__builtin_rvv_vle16_v_i16m1((const int16_t *)(op0), (size_t)(op1)) +#define vle16_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vle16_v_i16m2(op0, op1) \ +__builtin_rvv_vle16_v_i16m2((const int16_t *)(op0), (size_t)(op1)) +#define vle16_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vle16_v_i16m4(op0, op1) \ +__builtin_rvv_vle16_v_i16m4((const int16_t *)(op0), (size_t)(op1)) +#define vle16_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vle16_v_i16m8(op0, op1) \ +__builtin_rvv_vle16_v_i16m8((const int16_t *)(op0), (size_t)(op1)) +#define vle16_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vle16_v_i16mf2(op0, op1) \ +__builtin_rvv_vle16_v_i16mf2((const int16_t *)(op0), (size_t)(op1)) +#define vle16_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vle16_v_i16mf4(op0, op1) \ +__builtin_rvv_vle16_v_i16mf4((const int16_t *)(op0), (size_t)(op1)) +#define vle16_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vle16_v_u16m1(op0, op1) \ +__builtin_rvv_vle16_v_u16m1((const uint16_t *)(op0), (size_t)(op1)) +#define vle16_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vle16_v_u16m2(op0, op1) \ +__builtin_rvv_vle16_v_u16m2((const uint16_t *)(op0), (size_t)(op1)) +#define vle16_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vle16_v_u16m4(op0, op1) \ +__builtin_rvv_vle16_v_u16m4((const uint16_t *)(op0), (size_t)(op1)) +#define vle16_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vle16_v_u16m8(op0, op1) \ +__builtin_rvv_vle16_v_u16m8((const uint16_t *)(op0), (size_t)(op1)) +#define vle16_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vle16_v_u16mf2(op0, op1) \ +__builtin_rvv_vle16_v_u16mf2((const uint16_t *)(op0), (size_t)(op1)) +#define vle16_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vle16_v_u16mf4(op0, op1) \ +__builtin_rvv_vle16_v_u16mf4((const uint16_t *)(op0), (size_t)(op1)) +#define vle16_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vle32_v_i32m1(op0, op1) \ +__builtin_rvv_vle32_v_i32m1((const int32_t *)(op0), (size_t)(op1)) +#define vle32_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vle32_v_i32m2(op0, op1) \ +__builtin_rvv_vle32_v_i32m2((const int32_t *)(op0), (size_t)(op1)) +#define vle32_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vle32_v_i32m4(op0, op1) \ +__builtin_rvv_vle32_v_i32m4((const int32_t *)(op0), (size_t)(op1)) +#define vle32_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vle32_v_i32m8(op0, op1) \ +__builtin_rvv_vle32_v_i32m8((const int32_t *)(op0), (size_t)(op1)) +#define vle32_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vle32_v_i32mf2(op0, op1) \ +__builtin_rvv_vle32_v_i32mf2((const int32_t *)(op0), (size_t)(op1)) +#define vle32_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vle32_v_u32m1(op0, op1) \ +__builtin_rvv_vle32_v_u32m1((const uint32_t *)(op0), (size_t)(op1)) +#define vle32_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vle32_v_u32m2(op0, op1) \ +__builtin_rvv_vle32_v_u32m2((const uint32_t *)(op0), (size_t)(op1)) +#define vle32_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vle32_v_u32m4(op0, op1) \ +__builtin_rvv_vle32_v_u32m4((const uint32_t *)(op0), (size_t)(op1)) +#define vle32_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vle32_v_u32m8(op0, op1) \ +__builtin_rvv_vle32_v_u32m8((const uint32_t *)(op0), (size_t)(op1)) +#define vle32_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vle32_v_u32mf2(op0, op1) \ +__builtin_rvv_vle32_v_u32mf2((const uint32_t *)(op0), (size_t)(op1)) +#define vle32_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vlse8_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8m1((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8m2((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8m4((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_u8m8(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8m8((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8mf2((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8mf4((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8mf8((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vle64_v_i64m1(op0, op1) \ +__builtin_rvv_vle64_v_i64m1((const int64_t *)(op0), (size_t)(op1)) +#define vle64_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vle64_v_i64m2(op0, op1) \ +__builtin_rvv_vle64_v_i64m2((const int64_t *)(op0), (size_t)(op1)) +#define vle64_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vle64_v_i64m4(op0, op1) \ +__builtin_rvv_vle64_v_i64m4((const int64_t *)(op0), (size_t)(op1)) +#define vle64_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vle64_v_i64m8(op0, op1) \ +__builtin_rvv_vle64_v_i64m8((const int64_t *)(op0), (size_t)(op1)) +#define vle64_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vle64_v_u64m1(op0, op1) \ +__builtin_rvv_vle64_v_u64m1((const uint64_t *)(op0), (size_t)(op1)) +#define vle64_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vle64_v_u64m2(op0, op1) \ +__builtin_rvv_vle64_v_u64m2((const uint64_t *)(op0), (size_t)(op1)) +#define vle64_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vle64_v_u64m4(op0, op1) \ +__builtin_rvv_vle64_v_u64m4((const uint64_t *)(op0), (size_t)(op1)) +#define vle64_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vle64_v_u64m8(op0, op1) \ +__builtin_rvv_vle64_v_u64m8((const uint64_t *)(op0), (size_t)(op1)) +#define vle64_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vse16_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vse16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vse16_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vse16_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vse16_v_i16m2((int16_t *)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vse16_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vse16_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vse16_v_i16m4((int16_t *)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vse16_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vse16_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vse16_v_i16m8((int16_t *)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vse16_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_i16m8_m((vbool2_t)(op0), (int16_t *)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vse16_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vse16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vse16_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vse16_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vse16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vse16_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vse16_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vse16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vse16_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vse16_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vse16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vse16_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vse16_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vse16_v_u16m4((uint16_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vse16_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vse16_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vse16_v_u16m8((uint16_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vse16_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_u16m8_m((vbool2_t)(op0), (uint16_t *)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vse16_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vse16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vse16_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vse16_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vse16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vse16_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vse32_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vse32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vse32_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vse32_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vse32_v_i32m2((int32_t *)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vse32_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vse32_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vse32_v_i32m4((int32_t *)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vse32_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vse32_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vse32_v_i32m8((int32_t *)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vse32_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vse32_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vse32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vse32_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vse32_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vse32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vse32_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vse32_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vse32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vse32_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vse32_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vse32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vse32_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vse32_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vse32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vse32_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vse32_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vse32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vse32_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vluxei8_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8m1((const int8_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8m2((const int8_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8m4((const int8_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxei8_v_i8m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8m8((const int8_t *)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (const int8_t *)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vluxei8_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8mf2((const int8_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8mf4((const int8_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8mf8((const int8_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vse64_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vse64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vse64_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vse64_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vse64_v_i64m2((int64_t *)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vse64_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vse64_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vse64_v_i64m4((int64_t *)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vse64_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vse64_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vse64_v_i64m8((int64_t *)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vse64_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vse64_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vse64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vse64_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vse64_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vse64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vse64_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vse64_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vse64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vse64_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vse64_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vse64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vse64_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vlse16_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vlse16_v_i16m1((const int16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vlse16_v_i16m2((const int16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vlse16_v_i16m4((const int16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vlse16_v_i16m8((const int16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vlse16_v_i16mf2((const int16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vlse16_v_i16mf4((const int16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vlse16_v_u16m1((const uint16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vlse16_v_u16m2((const uint16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vlse16_v_u16m4((const uint16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vlse16_v_u16m8((const uint16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vlse16_v_u16mf2((const uint16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vlse16_v_u16mf4((const uint16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vlse32_v_i32m1((const int32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vlse32_v_i32m2((const int32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vlse32_v_i32m4((const int32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vlse32_v_i32m8((const int32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vlse32_v_i32mf2((const int32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vlse32_v_u32m1((const uint32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vlse32_v_u32m2((const uint32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vlse32_v_u32m4((const uint32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vlse32_v_u32m8((const uint32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vlse32_v_u32mf2((const uint32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vlse64_v_i64m1((const int64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vlse64_v_i64m2((const int64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vlse64_v_i64m4((const int64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vlse64_v_i64m8((const int64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vfirst_m_b8(op0, op1) \ +__builtin_rvv_vfirst_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vfirst_m_b8_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vfirst_m_b4(op0, op1) \ +__builtin_rvv_vfirst_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vfirst_m_b4_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vfirst_m_b2(op0, op1) \ +__builtin_rvv_vfirst_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vfirst_m_b2_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vfirst_m_b1(op0, op1) \ +__builtin_rvv_vfirst_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vfirst_m_b1_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vfirst_m_b16(op0, op1) \ +__builtin_rvv_vfirst_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vfirst_m_b16_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vfirst_m_b32(op0, op1) \ +__builtin_rvv_vfirst_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vfirst_m_b32_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vfirst_m_b64(op0, op1) \ +__builtin_rvv_vfirst_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define vfirst_m_b64_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vle1_v_b8(op0, op1) \ +__builtin_rvv_vle1_v_b8((const uint8_t *)(op0), (size_t)(op1)) +#define vle1_v_b4(op0, op1) \ +__builtin_rvv_vle1_v_b4((const uint8_t *)(op0), (size_t)(op1)) +#define vle1_v_b2(op0, op1) \ +__builtin_rvv_vle1_v_b2((const uint8_t *)(op0), (size_t)(op1)) +#define vle1_v_b1(op0, op1) \ +__builtin_rvv_vle1_v_b1((const uint8_t *)(op0), (size_t)(op1)) +#define vle1_v_b16(op0, op1) \ +__builtin_rvv_vle1_v_b16((const uint8_t *)(op0), (size_t)(op1)) +#define vle1_v_b32(op0, op1) \ +__builtin_rvv_vle1_v_b32((const uint8_t *)(op0), (size_t)(op1)) +#define vle1_v_b64(op0, op1) \ +__builtin_rvv_vle1_v_b64((const uint8_t *)(op0), (size_t)(op1)) +#define vlmul_ext_v_i8mf2_i8m1(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf2_i8m1((vint8mf2_t)(op0)) +#define vlmul_ext_v_i8mf4_i8m1(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf4_i8m1((vint8mf4_t)(op0)) +#define vlmul_ext_v_i8mf8_i8m1(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf8_i8m1((vint8mf8_t)(op0)) +#define vlmul_ext_v_i16mf2_i16m1(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf2_i16m1((vint16mf2_t)(op0)) +#define vlmul_ext_v_i16mf4_i16m1(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf4_i16m1((vint16mf4_t)(op0)) +#define vlmul_ext_v_i32mf2_i32m1(op0) \ +__builtin_rvv_vlmul_ext_v_i32mf2_i32m1((vint32mf2_t)(op0)) +#define vlmul_ext_v_i8mf4_i8mf2(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf4_i8mf2((vint8mf4_t)(op0)) +#define vlmul_ext_v_i8mf8_i8mf2(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf8_i8mf2((vint8mf8_t)(op0)) +#define vlmul_ext_v_i16mf4_i16mf2(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf4_i16mf2((vint16mf4_t)(op0)) +#define vlmul_ext_v_i8mf8_i8mf4(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf8_i8mf4((vint8mf8_t)(op0)) +#define vlmul_ext_v_i8m1_i8m2(op0) \ +__builtin_rvv_vlmul_ext_v_i8m1_i8m2((vint8m1_t)(op0)) +#define vlmul_ext_v_i8mf2_i8m2(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf2_i8m2((vint8mf2_t)(op0)) +#define vlmul_ext_v_i8mf4_i8m2(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf4_i8m2((vint8mf4_t)(op0)) +#define vlmul_ext_v_i8mf8_i8m2(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf8_i8m2((vint8mf8_t)(op0)) +#define vlmul_ext_v_i16m1_i16m2(op0) \ +__builtin_rvv_vlmul_ext_v_i16m1_i16m2((vint16m1_t)(op0)) +#define vlmul_ext_v_i16mf2_i16m2(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf2_i16m2((vint16mf2_t)(op0)) +#define vlmul_ext_v_i16mf4_i16m2(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf4_i16m2((vint16mf4_t)(op0)) +#define vlmul_ext_v_i32m1_i32m2(op0) \ +__builtin_rvv_vlmul_ext_v_i32m1_i32m2((vint32m1_t)(op0)) +#define vlmul_ext_v_i32mf2_i32m2(op0) \ +__builtin_rvv_vlmul_ext_v_i32mf2_i32m2((vint32mf2_t)(op0)) +#define vlmul_ext_v_i64m1_i64m2(op0) \ +__builtin_rvv_vlmul_ext_v_i64m1_i64m2((vint64m1_t)(op0)) +#define vlmul_ext_v_i8m1_i8m4(op0) \ +__builtin_rvv_vlmul_ext_v_i8m1_i8m4((vint8m1_t)(op0)) +#define vlmul_ext_v_i8m2_i8m4(op0) \ +__builtin_rvv_vlmul_ext_v_i8m2_i8m4((vint8m2_t)(op0)) +#define vlmul_ext_v_i8mf2_i8m4(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf2_i8m4((vint8mf2_t)(op0)) +#define vlmul_ext_v_i8mf4_i8m4(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf4_i8m4((vint8mf4_t)(op0)) +#define vlmul_ext_v_i8mf8_i8m4(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf8_i8m4((vint8mf8_t)(op0)) +#define vlmul_ext_v_i16m1_i16m4(op0) \ +__builtin_rvv_vlmul_ext_v_i16m1_i16m4((vint16m1_t)(op0)) +#define vlmul_ext_v_i16m2_i16m4(op0) \ +__builtin_rvv_vlmul_ext_v_i16m2_i16m4((vint16m2_t)(op0)) +#define vlmul_ext_v_i16mf2_i16m4(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf2_i16m4((vint16mf2_t)(op0)) +#define vlmul_ext_v_i16mf4_i16m4(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf4_i16m4((vint16mf4_t)(op0)) +#define vlmul_ext_v_i32m1_i32m4(op0) \ +__builtin_rvv_vlmul_ext_v_i32m1_i32m4((vint32m1_t)(op0)) +#define vlmul_ext_v_i32m2_i32m4(op0) \ +__builtin_rvv_vlmul_ext_v_i32m2_i32m4((vint32m2_t)(op0)) +#define vlmul_ext_v_i32mf2_i32m4(op0) \ +__builtin_rvv_vlmul_ext_v_i32mf2_i32m4((vint32mf2_t)(op0)) +#define vlmul_ext_v_i64m1_i64m4(op0) \ +__builtin_rvv_vlmul_ext_v_i64m1_i64m4((vint64m1_t)(op0)) +#define vlmul_ext_v_i64m2_i64m4(op0) \ +__builtin_rvv_vlmul_ext_v_i64m2_i64m4((vint64m2_t)(op0)) +#define vlmul_ext_v_i8m1_i8m8(op0) \ +__builtin_rvv_vlmul_ext_v_i8m1_i8m8((vint8m1_t)(op0)) +#define vlmul_ext_v_i8m2_i8m8(op0) \ +__builtin_rvv_vlmul_ext_v_i8m2_i8m8((vint8m2_t)(op0)) +#define vlmul_ext_v_i8m4_i8m8(op0) \ +__builtin_rvv_vlmul_ext_v_i8m4_i8m8((vint8m4_t)(op0)) +#define vlmul_ext_v_i8mf2_i8m8(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf2_i8m8((vint8mf2_t)(op0)) +#define vlmul_ext_v_i8mf4_i8m8(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf4_i8m8((vint8mf4_t)(op0)) +#define vlmul_ext_v_i8mf8_i8m8(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf8_i8m8((vint8mf8_t)(op0)) +#define vlmul_ext_v_i16m1_i16m8(op0) \ +__builtin_rvv_vlmul_ext_v_i16m1_i16m8((vint16m1_t)(op0)) +#define vlmul_ext_v_i16m2_i16m8(op0) \ +__builtin_rvv_vlmul_ext_v_i16m2_i16m8((vint16m2_t)(op0)) +#define vlmul_ext_v_i16m4_i16m8(op0) \ +__builtin_rvv_vlmul_ext_v_i16m4_i16m8((vint16m4_t)(op0)) +#define vlmul_ext_v_i16mf2_i16m8(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf2_i16m8((vint16mf2_t)(op0)) +#define vlmul_ext_v_i16mf4_i16m8(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf4_i16m8((vint16mf4_t)(op0)) +#define vlmul_ext_v_i32m1_i32m8(op0) \ +__builtin_rvv_vlmul_ext_v_i32m1_i32m8((vint32m1_t)(op0)) +#define vlmul_ext_v_i32m2_i32m8(op0) \ +__builtin_rvv_vlmul_ext_v_i32m2_i32m8((vint32m2_t)(op0)) +#define vlmul_ext_v_i32m4_i32m8(op0) \ +__builtin_rvv_vlmul_ext_v_i32m4_i32m8((vint32m4_t)(op0)) +#define vlmul_ext_v_i32mf2_i32m8(op0) \ +__builtin_rvv_vlmul_ext_v_i32mf2_i32m8((vint32mf2_t)(op0)) +#define vlmul_ext_v_i64m1_i64m8(op0) \ +__builtin_rvv_vlmul_ext_v_i64m1_i64m8((vint64m1_t)(op0)) +#define vlmul_ext_v_i64m2_i64m8(op0) \ +__builtin_rvv_vlmul_ext_v_i64m2_i64m8((vint64m2_t)(op0)) +#define vlmul_ext_v_i64m4_i64m8(op0) \ +__builtin_rvv_vlmul_ext_v_i64m4_i64m8((vint64m4_t)(op0)) +#define vlmul_ext_v_u8mf2_u8m1(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf2_u8m1((vuint8mf2_t)(op0)) +#define vlmul_ext_v_u8mf4_u8m1(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf4_u8m1((vuint8mf4_t)(op0)) +#define vlmul_ext_v_u8mf8_u8m1(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf8_u8m1((vuint8mf8_t)(op0)) +#define vlmul_ext_v_u16mf2_u16m1(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf2_u16m1((vuint16mf2_t)(op0)) +#define vlmul_ext_v_u16mf4_u16m1(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf4_u16m1((vuint16mf4_t)(op0)) +#define vlmul_ext_v_u32mf2_u32m1(op0) \ +__builtin_rvv_vlmul_ext_v_u32mf2_u32m1((vuint32mf2_t)(op0)) +#define vlmul_ext_v_u8mf4_u8mf2(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf4_u8mf2((vuint8mf4_t)(op0)) +#define vlmul_ext_v_u8mf8_u8mf2(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf8_u8mf2((vuint8mf8_t)(op0)) +#define vlmul_ext_v_u16mf4_u16mf2(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf4_u16mf2((vuint16mf4_t)(op0)) +#define vlmul_ext_v_u8mf8_u8mf4(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf8_u8mf4((vuint8mf8_t)(op0)) +#define vlmul_ext_v_u8m1_u8m2(op0) \ +__builtin_rvv_vlmul_ext_v_u8m1_u8m2((vuint8m1_t)(op0)) +#define vlmul_ext_v_u8mf2_u8m2(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf2_u8m2((vuint8mf2_t)(op0)) +#define vlmul_ext_v_u8mf4_u8m2(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf4_u8m2((vuint8mf4_t)(op0)) +#define vlmul_ext_v_u8mf8_u8m2(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf8_u8m2((vuint8mf8_t)(op0)) +#define vlmul_ext_v_u16m1_u16m2(op0) \ +__builtin_rvv_vlmul_ext_v_u16m1_u16m2((vuint16m1_t)(op0)) +#define vlmul_ext_v_u16mf2_u16m2(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf2_u16m2((vuint16mf2_t)(op0)) +#define vlmul_ext_v_u16mf4_u16m2(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf4_u16m2((vuint16mf4_t)(op0)) +#define vlmul_ext_v_u32m1_u32m2(op0) \ +__builtin_rvv_vlmul_ext_v_u32m1_u32m2((vuint32m1_t)(op0)) +#define vlmul_ext_v_u32mf2_u32m2(op0) \ +__builtin_rvv_vlmul_ext_v_u32mf2_u32m2((vuint32mf2_t)(op0)) +#define vlmul_ext_v_u64m1_u64m2(op0) \ +__builtin_rvv_vlmul_ext_v_u64m1_u64m2((vuint64m1_t)(op0)) +#define vlmul_ext_v_u8m1_u8m4(op0) \ +__builtin_rvv_vlmul_ext_v_u8m1_u8m4((vuint8m1_t)(op0)) +#define vlmul_ext_v_u8m2_u8m4(op0) \ +__builtin_rvv_vlmul_ext_v_u8m2_u8m4((vuint8m2_t)(op0)) +#define vlmul_ext_v_u8mf2_u8m4(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf2_u8m4((vuint8mf2_t)(op0)) +#define vlmul_ext_v_u8mf4_u8m4(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf4_u8m4((vuint8mf4_t)(op0)) +#define vlmul_ext_v_u8mf8_u8m4(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf8_u8m4((vuint8mf8_t)(op0)) +#define vlmul_ext_v_u16m1_u16m4(op0) \ +__builtin_rvv_vlmul_ext_v_u16m1_u16m4((vuint16m1_t)(op0)) +#define vlmul_ext_v_u16m2_u16m4(op0) \ +__builtin_rvv_vlmul_ext_v_u16m2_u16m4((vuint16m2_t)(op0)) +#define vlmul_ext_v_u16mf2_u16m4(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf2_u16m4((vuint16mf2_t)(op0)) +#define vlmul_ext_v_u16mf4_u16m4(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf4_u16m4((vuint16mf4_t)(op0)) +#define vlmul_ext_v_u32m1_u32m4(op0) \ +__builtin_rvv_vlmul_ext_v_u32m1_u32m4((vuint32m1_t)(op0)) +#define vlmul_ext_v_u32m2_u32m4(op0) \ +__builtin_rvv_vlmul_ext_v_u32m2_u32m4((vuint32m2_t)(op0)) +#define vlmul_ext_v_u32mf2_u32m4(op0) \ +__builtin_rvv_vlmul_ext_v_u32mf2_u32m4((vuint32mf2_t)(op0)) +#define vlmul_ext_v_u64m1_u64m4(op0) \ +__builtin_rvv_vlmul_ext_v_u64m1_u64m4((vuint64m1_t)(op0)) +#define vlmul_ext_v_u64m2_u64m4(op0) \ +__builtin_rvv_vlmul_ext_v_u64m2_u64m4((vuint64m2_t)(op0)) +#define vlmul_ext_v_u8m1_u8m8(op0) \ +__builtin_rvv_vlmul_ext_v_u8m1_u8m8((vuint8m1_t)(op0)) +#define vlmul_ext_v_u8m2_u8m8(op0) \ +__builtin_rvv_vlmul_ext_v_u8m2_u8m8((vuint8m2_t)(op0)) +#define vlmul_ext_v_u8m4_u8m8(op0) \ +__builtin_rvv_vlmul_ext_v_u8m4_u8m8((vuint8m4_t)(op0)) +#define vlmul_ext_v_u8mf2_u8m8(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf2_u8m8((vuint8mf2_t)(op0)) +#define vlmul_ext_v_u8mf4_u8m8(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf4_u8m8((vuint8mf4_t)(op0)) +#define vlmul_ext_v_u8mf8_u8m8(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf8_u8m8((vuint8mf8_t)(op0)) +#define vlmul_ext_v_u16m1_u16m8(op0) \ +__builtin_rvv_vlmul_ext_v_u16m1_u16m8((vuint16m1_t)(op0)) +#define vlmul_ext_v_u16m2_u16m8(op0) \ +__builtin_rvv_vlmul_ext_v_u16m2_u16m8((vuint16m2_t)(op0)) +#define vlmul_ext_v_u16m4_u16m8(op0) \ +__builtin_rvv_vlmul_ext_v_u16m4_u16m8((vuint16m4_t)(op0)) +#define vlmul_ext_v_u16mf2_u16m8(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf2_u16m8((vuint16mf2_t)(op0)) +#define vlmul_ext_v_u16mf4_u16m8(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf4_u16m8((vuint16mf4_t)(op0)) +#define vlmul_ext_v_u32m1_u32m8(op0) \ +__builtin_rvv_vlmul_ext_v_u32m1_u32m8((vuint32m1_t)(op0)) +#define vlmul_ext_v_u32m2_u32m8(op0) \ +__builtin_rvv_vlmul_ext_v_u32m2_u32m8((vuint32m2_t)(op0)) +#define vlmul_ext_v_u32m4_u32m8(op0) \ +__builtin_rvv_vlmul_ext_v_u32m4_u32m8((vuint32m4_t)(op0)) +#define vlmul_ext_v_u32mf2_u32m8(op0) \ +__builtin_rvv_vlmul_ext_v_u32mf2_u32m8((vuint32mf2_t)(op0)) +#define vlmul_ext_v_u64m1_u64m8(op0) \ +__builtin_rvv_vlmul_ext_v_u64m1_u64m8((vuint64m1_t)(op0)) +#define vlmul_ext_v_u64m2_u64m8(op0) \ +__builtin_rvv_vlmul_ext_v_u64m2_u64m8((vuint64m2_t)(op0)) +#define vlmul_ext_v_u64m4_u64m8(op0) \ +__builtin_rvv_vlmul_ext_v_u64m4_u64m8((vuint64m4_t)(op0)) +#define vlmul_trunc_v_i8m1_i8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m1_i8mf2((vint8m1_t)(op0)) +#define vlmul_trunc_v_i8m2_i8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m2_i8mf2((vint8m2_t)(op0)) +#define vlmul_trunc_v_i8m4_i8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m4_i8mf2((vint8m4_t)(op0)) +#define vlmul_trunc_v_i8m8_i8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m8_i8mf2((vint8m8_t)(op0)) +#define vlmul_trunc_v_i16m1_i16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m1_i16mf2((vint16m1_t)(op0)) +#define vlmul_trunc_v_i16m2_i16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m2_i16mf2((vint16m2_t)(op0)) +#define vlmul_trunc_v_i16m4_i16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m4_i16mf2((vint16m4_t)(op0)) +#define vlmul_trunc_v_i16m8_i16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m8_i16mf2((vint16m8_t)(op0)) +#define vlmul_trunc_v_i32m1_i32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m1_i32mf2((vint32m1_t)(op0)) +#define vlmul_trunc_v_i32m2_i32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m2_i32mf2((vint32m2_t)(op0)) +#define vlmul_trunc_v_i32m4_i32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m4_i32mf2((vint32m4_t)(op0)) +#define vlmul_trunc_v_i32m8_i32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m8_i32mf2((vint32m8_t)(op0)) +#define vlmul_trunc_v_i8m1_i8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m1_i8mf4((vint8m1_t)(op0)) +#define vlmul_trunc_v_i8m2_i8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m2_i8mf4((vint8m2_t)(op0)) +#define vlmul_trunc_v_i8m4_i8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m4_i8mf4((vint8m4_t)(op0)) +#define vlmul_trunc_v_i8m8_i8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m8_i8mf4((vint8m8_t)(op0)) +#define vlmul_trunc_v_i8mf2_i8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i8mf2_i8mf4((vint8mf2_t)(op0)) +#define vlmul_trunc_v_i16m1_i16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m1_i16mf4((vint16m1_t)(op0)) +#define vlmul_trunc_v_i16m2_i16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m2_i16mf4((vint16m2_t)(op0)) +#define vlmul_trunc_v_i16m4_i16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m4_i16mf4((vint16m4_t)(op0)) +#define vlmul_trunc_v_i16m8_i16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m8_i16mf4((vint16m8_t)(op0)) +#define vlmul_trunc_v_i16mf2_i16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i16mf2_i16mf4((vint16mf2_t)(op0)) +#define vlmul_trunc_v_i8m1_i8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m1_i8mf8((vint8m1_t)(op0)) +#define vlmul_trunc_v_i8m2_i8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m2_i8mf8((vint8m2_t)(op0)) +#define vlmul_trunc_v_i8m4_i8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m4_i8mf8((vint8m4_t)(op0)) +#define vlmul_trunc_v_i8m8_i8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m8_i8mf8((vint8m8_t)(op0)) +#define vlmul_trunc_v_i8mf2_i8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_i8mf2_i8mf8((vint8mf2_t)(op0)) +#define vlmul_trunc_v_i8mf4_i8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_i8mf4_i8mf8((vint8mf4_t)(op0)) +#define vlmul_trunc_v_i8m2_i8m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m2_i8m1((vint8m2_t)(op0)) +#define vlmul_trunc_v_i8m4_i8m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m4_i8m1((vint8m4_t)(op0)) +#define vlmul_trunc_v_i8m8_i8m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m8_i8m1((vint8m8_t)(op0)) +#define vlmul_trunc_v_i16m2_i16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m2_i16m1((vint16m2_t)(op0)) +#define vlmul_trunc_v_i16m4_i16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m4_i16m1((vint16m4_t)(op0)) +#define vlmul_trunc_v_i16m8_i16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m8_i16m1((vint16m8_t)(op0)) +#define vlmul_trunc_v_i32m2_i32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m2_i32m1((vint32m2_t)(op0)) +#define vlmul_trunc_v_i32m4_i32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m4_i32m1((vint32m4_t)(op0)) +#define vlmul_trunc_v_i32m8_i32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m8_i32m1((vint32m8_t)(op0)) +#define vlmul_trunc_v_i64m2_i64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i64m2_i64m1((vint64m2_t)(op0)) +#define vlmul_trunc_v_i64m4_i64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i64m4_i64m1((vint64m4_t)(op0)) +#define vlmul_trunc_v_i64m8_i64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i64m8_i64m1((vint64m8_t)(op0)) +#define vlmul_trunc_v_i8m4_i8m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m4_i8m2((vint8m4_t)(op0)) +#define vlmul_trunc_v_i8m8_i8m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m8_i8m2((vint8m8_t)(op0)) +#define vlmul_trunc_v_i16m4_i16m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m4_i16m2((vint16m4_t)(op0)) +#define vlmul_trunc_v_i16m8_i16m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m8_i16m2((vint16m8_t)(op0)) +#define vlmul_trunc_v_i32m4_i32m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m4_i32m2((vint32m4_t)(op0)) +#define vlmul_trunc_v_i32m8_i32m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m8_i32m2((vint32m8_t)(op0)) +#define vlmul_trunc_v_i64m4_i64m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i64m4_i64m2((vint64m4_t)(op0)) +#define vlmul_trunc_v_i64m8_i64m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i64m8_i64m2((vint64m8_t)(op0)) +#define vlmul_trunc_v_i8m8_i8m4(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m8_i8m4((vint8m8_t)(op0)) +#define vlmul_trunc_v_i16m8_i16m4(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m8_i16m4((vint16m8_t)(op0)) +#define vlmul_trunc_v_i32m8_i32m4(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m8_i32m4((vint32m8_t)(op0)) +#define vlmul_trunc_v_i64m8_i64m4(op0) \ +__builtin_rvv_vlmul_trunc_v_i64m8_i64m4((vint64m8_t)(op0)) +#define vlmul_trunc_v_u8m1_u8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m1_u8mf2((vuint8m1_t)(op0)) +#define vlmul_trunc_v_u8m2_u8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m2_u8mf2((vuint8m2_t)(op0)) +#define vlmul_trunc_v_u8m4_u8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m4_u8mf2((vuint8m4_t)(op0)) +#define vlmul_trunc_v_u8m8_u8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m8_u8mf2((vuint8m8_t)(op0)) +#define vlmul_trunc_v_u16m1_u16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m1_u16mf2((vuint16m1_t)(op0)) +#define vlmul_trunc_v_u16m2_u16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m2_u16mf2((vuint16m2_t)(op0)) +#define vlmul_trunc_v_u16m4_u16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m4_u16mf2((vuint16m4_t)(op0)) +#define vlmul_trunc_v_u16m8_u16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m8_u16mf2((vuint16m8_t)(op0)) +#define vlmul_trunc_v_u32m1_u32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m1_u32mf2((vuint32m1_t)(op0)) +#define vlmul_trunc_v_u32m2_u32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m2_u32mf2((vuint32m2_t)(op0)) +#define vlmul_trunc_v_u32m4_u32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m4_u32mf2((vuint32m4_t)(op0)) +#define vlmul_trunc_v_u32m8_u32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m8_u32mf2((vuint32m8_t)(op0)) +#define vlmul_trunc_v_u8m1_u8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m1_u8mf4((vuint8m1_t)(op0)) +#define vlmul_trunc_v_u8m2_u8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m2_u8mf4((vuint8m2_t)(op0)) +#define vlmul_trunc_v_u8m4_u8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m4_u8mf4((vuint8m4_t)(op0)) +#define vlmul_trunc_v_u8m8_u8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m8_u8mf4((vuint8m8_t)(op0)) +#define vlmul_trunc_v_u8mf2_u8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u8mf2_u8mf4((vuint8mf2_t)(op0)) +#define vlmul_trunc_v_u16m1_u16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m1_u16mf4((vuint16m1_t)(op0)) +#define vlmul_trunc_v_u16m2_u16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m2_u16mf4((vuint16m2_t)(op0)) +#define vlmul_trunc_v_u16m4_u16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m4_u16mf4((vuint16m4_t)(op0)) +#define vlmul_trunc_v_u16m8_u16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m8_u16mf4((vuint16m8_t)(op0)) +#define vlmul_trunc_v_u16mf2_u16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u16mf2_u16mf4((vuint16mf2_t)(op0)) +#define vlmul_trunc_v_u8m1_u8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m1_u8mf8((vuint8m1_t)(op0)) +#define vlmul_trunc_v_u8m2_u8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m2_u8mf8((vuint8m2_t)(op0)) +#define vlmul_trunc_v_u8m4_u8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m4_u8mf8((vuint8m4_t)(op0)) +#define vlmul_trunc_v_u8m8_u8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m8_u8mf8((vuint8m8_t)(op0)) +#define vlmul_trunc_v_u8mf2_u8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_u8mf2_u8mf8((vuint8mf2_t)(op0)) +#define vlmul_trunc_v_u8mf4_u8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_u8mf4_u8mf8((vuint8mf4_t)(op0)) +#define vlmul_trunc_v_u8m2_u8m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m2_u8m1((vuint8m2_t)(op0)) +#define vlmul_trunc_v_u8m4_u8m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m4_u8m1((vuint8m4_t)(op0)) +#define vlmul_trunc_v_u8m8_u8m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m8_u8m1((vuint8m8_t)(op0)) +#define vlmul_trunc_v_u16m2_u16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m2_u16m1((vuint16m2_t)(op0)) +#define vlmul_trunc_v_u16m4_u16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m4_u16m1((vuint16m4_t)(op0)) +#define vlmul_trunc_v_u16m8_u16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m8_u16m1((vuint16m8_t)(op0)) +#define vlmul_trunc_v_u32m2_u32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m2_u32m1((vuint32m2_t)(op0)) +#define vlmul_trunc_v_u32m4_u32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m4_u32m1((vuint32m4_t)(op0)) +#define vlmul_trunc_v_u32m8_u32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m8_u32m1((vuint32m8_t)(op0)) +#define vlmul_trunc_v_u64m2_u64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u64m2_u64m1((vuint64m2_t)(op0)) +#define vlmul_trunc_v_u64m4_u64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u64m4_u64m1((vuint64m4_t)(op0)) +#define vlmul_trunc_v_u64m8_u64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u64m8_u64m1((vuint64m8_t)(op0)) +#define vlmul_trunc_v_u8m4_u8m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m4_u8m2((vuint8m4_t)(op0)) +#define vlmul_trunc_v_u8m8_u8m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m8_u8m2((vuint8m8_t)(op0)) +#define vlmul_trunc_v_u16m4_u16m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m4_u16m2((vuint16m4_t)(op0)) +#define vlmul_trunc_v_u16m8_u16m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m8_u16m2((vuint16m8_t)(op0)) +#define vlmul_trunc_v_u32m4_u32m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m4_u32m2((vuint32m4_t)(op0)) +#define vlmul_trunc_v_u32m8_u32m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m8_u32m2((vuint32m8_t)(op0)) +#define vlmul_trunc_v_u64m4_u64m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u64m4_u64m2((vuint64m4_t)(op0)) +#define vlmul_trunc_v_u64m8_u64m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u64m8_u64m2((vuint64m8_t)(op0)) +#define vlmul_trunc_v_u8m8_u8m4(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m8_u8m4((vuint8m8_t)(op0)) +#define vlmul_trunc_v_u16m8_u16m4(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m8_u16m4((vuint16m8_t)(op0)) +#define vlmul_trunc_v_u32m8_u32m4(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m8_u32m4((vuint32m8_t)(op0)) +#define vlmul_trunc_v_u64m8_u64m4(op0) \ +__builtin_rvv_vlmul_trunc_v_u64m8_u64m4((vuint64m8_t)(op0)) +#define vmand_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmand_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmand_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmand_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmand_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmand_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmand_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmclr_m_b8(op0) \ +__builtin_rvv_vmclr_m_b8((size_t)(op0)) +#define vmclr_m_b4(op0) \ +__builtin_rvv_vmclr_m_b4((size_t)(op0)) +#define vmclr_m_b2(op0) \ +__builtin_rvv_vmclr_m_b2((size_t)(op0)) +#define vmclr_m_b1(op0) \ +__builtin_rvv_vmclr_m_b1((size_t)(op0)) +#define vmclr_m_b16(op0) \ +__builtin_rvv_vmclr_m_b16((size_t)(op0)) +#define vmclr_m_b32(op0) \ +__builtin_rvv_vmclr_m_b32((size_t)(op0)) +#define vmclr_m_b64(op0) \ +__builtin_rvv_vmclr_m_b64((size_t)(op0)) +#define vmnand_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmnand_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmnand_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmnand_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmnand_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmnand_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmnand_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmnor_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmnor_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmnor_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmnor_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmnor_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmnor_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmnor_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmor_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmor_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmor_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmor_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmor_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmor_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmor_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmornot_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmornot_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmornot_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmornot_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmornot_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmornot_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmornot_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmsbf_m_b8(op0, op1) \ +__builtin_rvv_vmsbf_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vmsbf_m_b8_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbf_m_b4(op0, op1) \ +__builtin_rvv_vmsbf_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vmsbf_m_b4_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbf_m_b2(op0, op1) \ +__builtin_rvv_vmsbf_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vmsbf_m_b2_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbf_m_b1(op0, op1) \ +__builtin_rvv_vmsbf_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vmsbf_m_b1_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsbf_m_b16(op0, op1) \ +__builtin_rvv_vmsbf_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vmsbf_m_b16_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbf_m_b32(op0, op1) \ +__builtin_rvv_vmsbf_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vmsbf_m_b32_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbf_m_b64(op0, op1) \ +__builtin_rvv_vmsbf_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define vmsbf_m_b64_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmset_m_b8(op0) \ +__builtin_rvv_vmset_m_b8((size_t)(op0)) +#define vmset_m_b4(op0) \ +__builtin_rvv_vmset_m_b4((size_t)(op0)) +#define vmset_m_b2(op0) \ +__builtin_rvv_vmset_m_b2((size_t)(op0)) +#define vmset_m_b1(op0) \ +__builtin_rvv_vmset_m_b1((size_t)(op0)) +#define vmset_m_b16(op0) \ +__builtin_rvv_vmset_m_b16((size_t)(op0)) +#define vmset_m_b32(op0) \ +__builtin_rvv_vmset_m_b32((size_t)(op0)) +#define vmset_m_b64(op0) \ +__builtin_rvv_vmset_m_b64((size_t)(op0)) +#define vmsif_m_b8(op0, op1) \ +__builtin_rvv_vmsif_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vmsif_m_b8_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsif_m_b4(op0, op1) \ +__builtin_rvv_vmsif_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vmsif_m_b4_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsif_m_b2(op0, op1) \ +__builtin_rvv_vmsif_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vmsif_m_b2_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsif_m_b1(op0, op1) \ +__builtin_rvv_vmsif_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vmsif_m_b1_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsif_m_b16(op0, op1) \ +__builtin_rvv_vmsif_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vmsif_m_b16_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsif_m_b32(op0, op1) \ +__builtin_rvv_vmsif_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vmsif_m_b32_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsif_m_b64(op0, op1) \ +__builtin_rvv_vmsif_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define vmsif_m_b64_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsof_m_b8(op0, op1) \ +__builtin_rvv_vmsof_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vmsof_m_b8_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsof_m_b4(op0, op1) \ +__builtin_rvv_vmsof_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vmsof_m_b4_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsof_m_b2(op0, op1) \ +__builtin_rvv_vmsof_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vmsof_m_b2_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsof_m_b1(op0, op1) \ +__builtin_rvv_vmsof_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vmsof_m_b1_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsof_m_b16(op0, op1) \ +__builtin_rvv_vmsof_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vmsof_m_b16_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsof_m_b32(op0, op1) \ +__builtin_rvv_vmsof_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vmsof_m_b32_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsof_m_b64(op0, op1) \ +__builtin_rvv_vmsof_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define vmsof_m_b64_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmxnor_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmxnor_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmxnor_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmxnor_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmxnor_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmxnor_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmxnor_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmxor_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmxor_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmxor_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmxor_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmxor_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmxor_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmxor_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vpopc_m_b8(op0, op1) \ +__builtin_rvv_vpopc_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vpopc_m_b8_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vpopc_m_b4(op0, op1) \ +__builtin_rvv_vpopc_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vpopc_m_b4_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vpopc_m_b2(op0, op1) \ +__builtin_rvv_vpopc_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vpopc_m_b2_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vpopc_m_b1(op0, op1) \ +__builtin_rvv_vpopc_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vpopc_m_b1_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vpopc_m_b16(op0, op1) \ +__builtin_rvv_vpopc_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vpopc_m_b16_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vpopc_m_b32(op0, op1) \ +__builtin_rvv_vpopc_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vpopc_m_b32_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vpopc_m_b64(op0, op1) \ +__builtin_rvv_vpopc_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define vpopc_m_b64_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vreinterpret_v_i8m1_i16m1(op0) \ +__builtin_rvv_vreinterpret_v_i8m1_i16m1((vint8m1_t)(op0)) +#define vreinterpret_v_i8m2_i16m2(op0) \ +__builtin_rvv_vreinterpret_v_i8m2_i16m2((vint8m2_t)(op0)) +#define vreinterpret_v_i8m4_i16m4(op0) \ +__builtin_rvv_vreinterpret_v_i8m4_i16m4((vint8m4_t)(op0)) +#define vreinterpret_v_i8m8_i16m8(op0) \ +__builtin_rvv_vreinterpret_v_i8m8_i16m8((vint8m8_t)(op0)) +#define vreinterpret_v_i8mf2_i16mf2(op0) \ +__builtin_rvv_vreinterpret_v_i8mf2_i16mf2((vint8mf2_t)(op0)) +#define vreinterpret_v_i8mf4_i16mf4(op0) \ +__builtin_rvv_vreinterpret_v_i8mf4_i16mf4((vint8mf4_t)(op0)) +#define vreinterpret_v_i32m1_i16m1(op0) \ +__builtin_rvv_vreinterpret_v_i32m1_i16m1((vint32m1_t)(op0)) +#define vreinterpret_v_i32m2_i16m2(op0) \ +__builtin_rvv_vreinterpret_v_i32m2_i16m2((vint32m2_t)(op0)) +#define vreinterpret_v_i32m4_i16m4(op0) \ +__builtin_rvv_vreinterpret_v_i32m4_i16m4((vint32m4_t)(op0)) +#define vreinterpret_v_i32m8_i16m8(op0) \ +__builtin_rvv_vreinterpret_v_i32m8_i16m8((vint32m8_t)(op0)) +#define vreinterpret_v_i32mf2_i16mf2(op0) \ +__builtin_rvv_vreinterpret_v_i32mf2_i16mf2((vint32mf2_t)(op0)) +#define vreinterpret_v_i64m1_i16m1(op0) \ +__builtin_rvv_vreinterpret_v_i64m1_i16m1((vint64m1_t)(op0)) +#define vreinterpret_v_i64m2_i16m2(op0) \ +__builtin_rvv_vreinterpret_v_i64m2_i16m2((vint64m2_t)(op0)) +#define vreinterpret_v_i64m4_i16m4(op0) \ +__builtin_rvv_vreinterpret_v_i64m4_i16m4((vint64m4_t)(op0)) +#define vreinterpret_v_i64m8_i16m8(op0) \ +__builtin_rvv_vreinterpret_v_i64m8_i16m8((vint64m8_t)(op0)) +#define vreinterpret_v_i8m1_i32m1(op0) \ +__builtin_rvv_vreinterpret_v_i8m1_i32m1((vint8m1_t)(op0)) +#define vreinterpret_v_i8m2_i32m2(op0) \ +__builtin_rvv_vreinterpret_v_i8m2_i32m2((vint8m2_t)(op0)) +#define vreinterpret_v_i8m4_i32m4(op0) \ +__builtin_rvv_vreinterpret_v_i8m4_i32m4((vint8m4_t)(op0)) +#define vreinterpret_v_i8m8_i32m8(op0) \ +__builtin_rvv_vreinterpret_v_i8m8_i32m8((vint8m8_t)(op0)) +#define vreinterpret_v_i8mf2_i32mf2(op0) \ +__builtin_rvv_vreinterpret_v_i8mf2_i32mf2((vint8mf2_t)(op0)) +#define vreinterpret_v_i16m1_i32m1(op0) \ +__builtin_rvv_vreinterpret_v_i16m1_i32m1((vint16m1_t)(op0)) +#define vreinterpret_v_i16m2_i32m2(op0) \ +__builtin_rvv_vreinterpret_v_i16m2_i32m2((vint16m2_t)(op0)) +#define vreinterpret_v_i16m4_i32m4(op0) \ +__builtin_rvv_vreinterpret_v_i16m4_i32m4((vint16m4_t)(op0)) +#define vreinterpret_v_i16m8_i32m8(op0) \ +__builtin_rvv_vreinterpret_v_i16m8_i32m8((vint16m8_t)(op0)) +#define vreinterpret_v_i16mf2_i32mf2(op0) \ +__builtin_rvv_vreinterpret_v_i16mf2_i32mf2((vint16mf2_t)(op0)) +#define vreinterpret_v_i64m1_i32m1(op0) \ +__builtin_rvv_vreinterpret_v_i64m1_i32m1((vint64m1_t)(op0)) +#define vreinterpret_v_i64m2_i32m2(op0) \ +__builtin_rvv_vreinterpret_v_i64m2_i32m2((vint64m2_t)(op0)) +#define vreinterpret_v_i64m4_i32m4(op0) \ +__builtin_rvv_vreinterpret_v_i64m4_i32m4((vint64m4_t)(op0)) +#define vreinterpret_v_i64m8_i32m8(op0) \ +__builtin_rvv_vreinterpret_v_i64m8_i32m8((vint64m8_t)(op0)) +#define vreinterpret_v_i8m1_i64m1(op0) \ +__builtin_rvv_vreinterpret_v_i8m1_i64m1((vint8m1_t)(op0)) +#define vreinterpret_v_i8m2_i64m2(op0) \ +__builtin_rvv_vreinterpret_v_i8m2_i64m2((vint8m2_t)(op0)) +#define vreinterpret_v_i8m4_i64m4(op0) \ +__builtin_rvv_vreinterpret_v_i8m4_i64m4((vint8m4_t)(op0)) +#define vreinterpret_v_i8m8_i64m8(op0) \ +__builtin_rvv_vreinterpret_v_i8m8_i64m8((vint8m8_t)(op0)) +#define vreinterpret_v_i16m1_i64m1(op0) \ +__builtin_rvv_vreinterpret_v_i16m1_i64m1((vint16m1_t)(op0)) +#define vreinterpret_v_i16m2_i64m2(op0) \ +__builtin_rvv_vreinterpret_v_i16m2_i64m2((vint16m2_t)(op0)) +#define vreinterpret_v_i16m4_i64m4(op0) \ +__builtin_rvv_vreinterpret_v_i16m4_i64m4((vint16m4_t)(op0)) +#define vreinterpret_v_i16m8_i64m8(op0) \ +__builtin_rvv_vreinterpret_v_i16m8_i64m8((vint16m8_t)(op0)) +#define vreinterpret_v_i32m1_i64m1(op0) \ +__builtin_rvv_vreinterpret_v_i32m1_i64m1((vint32m1_t)(op0)) +#define vreinterpret_v_i32m2_i64m2(op0) \ +__builtin_rvv_vreinterpret_v_i32m2_i64m2((vint32m2_t)(op0)) +#define vreinterpret_v_i32m4_i64m4(op0) \ +__builtin_rvv_vreinterpret_v_i32m4_i64m4((vint32m4_t)(op0)) +#define vreinterpret_v_i32m8_i64m8(op0) \ +__builtin_rvv_vreinterpret_v_i32m8_i64m8((vint32m8_t)(op0)) +#define vreinterpret_v_i16m1_i8m1(op0) \ +__builtin_rvv_vreinterpret_v_i16m1_i8m1((vint16m1_t)(op0)) +#define vreinterpret_v_i16m2_i8m2(op0) \ +__builtin_rvv_vreinterpret_v_i16m2_i8m2((vint16m2_t)(op0)) +#define vreinterpret_v_i16m4_i8m4(op0) \ +__builtin_rvv_vreinterpret_v_i16m4_i8m4((vint16m4_t)(op0)) +#define vreinterpret_v_i16m8_i8m8(op0) \ +__builtin_rvv_vreinterpret_v_i16m8_i8m8((vint16m8_t)(op0)) +#define vreinterpret_v_i16mf2_i8mf2(op0) \ +__builtin_rvv_vreinterpret_v_i16mf2_i8mf2((vint16mf2_t)(op0)) +#define vreinterpret_v_i16mf4_i8mf4(op0) \ +__builtin_rvv_vreinterpret_v_i16mf4_i8mf4((vint16mf4_t)(op0)) +#define vreinterpret_v_i32m1_i8m1(op0) \ +__builtin_rvv_vreinterpret_v_i32m1_i8m1((vint32m1_t)(op0)) +#define vreinterpret_v_i32m2_i8m2(op0) \ +__builtin_rvv_vreinterpret_v_i32m2_i8m2((vint32m2_t)(op0)) +#define vreinterpret_v_i32m4_i8m4(op0) \ +__builtin_rvv_vreinterpret_v_i32m4_i8m4((vint32m4_t)(op0)) +#define vreinterpret_v_i32m8_i8m8(op0) \ +__builtin_rvv_vreinterpret_v_i32m8_i8m8((vint32m8_t)(op0)) +#define vreinterpret_v_i32mf2_i8mf2(op0) \ +__builtin_rvv_vreinterpret_v_i32mf2_i8mf2((vint32mf2_t)(op0)) +#define vreinterpret_v_i64m1_i8m1(op0) \ +__builtin_rvv_vreinterpret_v_i64m1_i8m1((vint64m1_t)(op0)) +#define vreinterpret_v_i64m2_i8m2(op0) \ +__builtin_rvv_vreinterpret_v_i64m2_i8m2((vint64m2_t)(op0)) +#define vreinterpret_v_i64m4_i8m4(op0) \ +__builtin_rvv_vreinterpret_v_i64m4_i8m4((vint64m4_t)(op0)) +#define vreinterpret_v_i64m8_i8m8(op0) \ +__builtin_rvv_vreinterpret_v_i64m8_i8m8((vint64m8_t)(op0)) +#define vreinterpret_v_u8m1_i8m1(op0) \ +__builtin_rvv_vreinterpret_v_u8m1_i8m1((vuint8m1_t)(op0)) +#define vreinterpret_v_u8m2_i8m2(op0) \ +__builtin_rvv_vreinterpret_v_u8m2_i8m2((vuint8m2_t)(op0)) +#define vreinterpret_v_u8m4_i8m4(op0) \ +__builtin_rvv_vreinterpret_v_u8m4_i8m4((vuint8m4_t)(op0)) +#define vreinterpret_v_u8m8_i8m8(op0) \ +__builtin_rvv_vreinterpret_v_u8m8_i8m8((vuint8m8_t)(op0)) +#define vreinterpret_v_u8mf2_i8mf2(op0) \ +__builtin_rvv_vreinterpret_v_u8mf2_i8mf2((vuint8mf2_t)(op0)) +#define vreinterpret_v_u8mf4_i8mf4(op0) \ +__builtin_rvv_vreinterpret_v_u8mf4_i8mf4((vuint8mf4_t)(op0)) +#define vreinterpret_v_u8mf8_i8mf8(op0) \ +__builtin_rvv_vreinterpret_v_u8mf8_i8mf8((vuint8mf8_t)(op0)) +#define vreinterpret_v_u16m1_i16m1(op0) \ +__builtin_rvv_vreinterpret_v_u16m1_i16m1((vuint16m1_t)(op0)) +#define vreinterpret_v_u16m2_i16m2(op0) \ +__builtin_rvv_vreinterpret_v_u16m2_i16m2((vuint16m2_t)(op0)) +#define vreinterpret_v_u16m4_i16m4(op0) \ +__builtin_rvv_vreinterpret_v_u16m4_i16m4((vuint16m4_t)(op0)) +#define vreinterpret_v_u16m8_i16m8(op0) \ +__builtin_rvv_vreinterpret_v_u16m8_i16m8((vuint16m8_t)(op0)) +#define vreinterpret_v_u16mf2_i16mf2(op0) \ +__builtin_rvv_vreinterpret_v_u16mf2_i16mf2((vuint16mf2_t)(op0)) +#define vreinterpret_v_u16mf4_i16mf4(op0) \ +__builtin_rvv_vreinterpret_v_u16mf4_i16mf4((vuint16mf4_t)(op0)) +#define vreinterpret_v_u32m1_i32m1(op0) \ +__builtin_rvv_vreinterpret_v_u32m1_i32m1((vuint32m1_t)(op0)) +#define vreinterpret_v_u32m2_i32m2(op0) \ +__builtin_rvv_vreinterpret_v_u32m2_i32m2((vuint32m2_t)(op0)) +#define vreinterpret_v_u32m4_i32m4(op0) \ +__builtin_rvv_vreinterpret_v_u32m4_i32m4((vuint32m4_t)(op0)) +#define vreinterpret_v_u32m8_i32m8(op0) \ +__builtin_rvv_vreinterpret_v_u32m8_i32m8((vuint32m8_t)(op0)) +#define vreinterpret_v_u32mf2_i32mf2(op0) \ +__builtin_rvv_vreinterpret_v_u32mf2_i32mf2((vuint32mf2_t)(op0)) +#define vreinterpret_v_u64m1_i64m1(op0) \ +__builtin_rvv_vreinterpret_v_u64m1_i64m1((vuint64m1_t)(op0)) +#define vreinterpret_v_u64m2_i64m2(op0) \ +__builtin_rvv_vreinterpret_v_u64m2_i64m2((vuint64m2_t)(op0)) +#define vreinterpret_v_u64m4_i64m4(op0) \ +__builtin_rvv_vreinterpret_v_u64m4_i64m4((vuint64m4_t)(op0)) +#define vreinterpret_v_u64m8_i64m8(op0) \ +__builtin_rvv_vreinterpret_v_u64m8_i64m8((vuint64m8_t)(op0)) +#define vreinterpret_v_u8m1_u16m1(op0) \ +__builtin_rvv_vreinterpret_v_u8m1_u16m1((vuint8m1_t)(op0)) +#define vreinterpret_v_u8m2_u16m2(op0) \ +__builtin_rvv_vreinterpret_v_u8m2_u16m2((vuint8m2_t)(op0)) +#define vreinterpret_v_u8m4_u16m4(op0) \ +__builtin_rvv_vreinterpret_v_u8m4_u16m4((vuint8m4_t)(op0)) +#define vreinterpret_v_u8m8_u16m8(op0) \ +__builtin_rvv_vreinterpret_v_u8m8_u16m8((vuint8m8_t)(op0)) +#define vreinterpret_v_u8mf2_u16mf2(op0) \ +__builtin_rvv_vreinterpret_v_u8mf2_u16mf2((vuint8mf2_t)(op0)) +#define vreinterpret_v_u8mf4_u16mf4(op0) \ +__builtin_rvv_vreinterpret_v_u8mf4_u16mf4((vuint8mf4_t)(op0)) +#define vreinterpret_v_u32m1_u16m1(op0) \ +__builtin_rvv_vreinterpret_v_u32m1_u16m1((vuint32m1_t)(op0)) +#define vreinterpret_v_u32m2_u16m2(op0) \ +__builtin_rvv_vreinterpret_v_u32m2_u16m2((vuint32m2_t)(op0)) +#define vreinterpret_v_u32m4_u16m4(op0) \ +__builtin_rvv_vreinterpret_v_u32m4_u16m4((vuint32m4_t)(op0)) +#define vreinterpret_v_u32m8_u16m8(op0) \ +__builtin_rvv_vreinterpret_v_u32m8_u16m8((vuint32m8_t)(op0)) +#define vreinterpret_v_u32mf2_u16mf2(op0) \ +__builtin_rvv_vreinterpret_v_u32mf2_u16mf2((vuint32mf2_t)(op0)) +#define vreinterpret_v_u64m1_u16m1(op0) \ +__builtin_rvv_vreinterpret_v_u64m1_u16m1((vuint64m1_t)(op0)) +#define vreinterpret_v_u64m2_u16m2(op0) \ +__builtin_rvv_vreinterpret_v_u64m2_u16m2((vuint64m2_t)(op0)) +#define vreinterpret_v_u64m4_u16m4(op0) \ +__builtin_rvv_vreinterpret_v_u64m4_u16m4((vuint64m4_t)(op0)) +#define vreinterpret_v_u64m8_u16m8(op0) \ +__builtin_rvv_vreinterpret_v_u64m8_u16m8((vuint64m8_t)(op0)) +#define vreinterpret_v_u8m1_u32m1(op0) \ +__builtin_rvv_vreinterpret_v_u8m1_u32m1((vuint8m1_t)(op0)) +#define vreinterpret_v_u8m2_u32m2(op0) \ +__builtin_rvv_vreinterpret_v_u8m2_u32m2((vuint8m2_t)(op0)) +#define vreinterpret_v_u8m4_u32m4(op0) \ +__builtin_rvv_vreinterpret_v_u8m4_u32m4((vuint8m4_t)(op0)) +#define vreinterpret_v_u8m8_u32m8(op0) \ +__builtin_rvv_vreinterpret_v_u8m8_u32m8((vuint8m8_t)(op0)) +#define vreinterpret_v_u8mf2_u32mf2(op0) \ +__builtin_rvv_vreinterpret_v_u8mf2_u32mf2((vuint8mf2_t)(op0)) +#define vreinterpret_v_u16m1_u32m1(op0) \ +__builtin_rvv_vreinterpret_v_u16m1_u32m1((vuint16m1_t)(op0)) +#define vreinterpret_v_u16m2_u32m2(op0) \ +__builtin_rvv_vreinterpret_v_u16m2_u32m2((vuint16m2_t)(op0)) +#define vreinterpret_v_u16m4_u32m4(op0) \ +__builtin_rvv_vreinterpret_v_u16m4_u32m4((vuint16m4_t)(op0)) +#define vreinterpret_v_u16m8_u32m8(op0) \ +__builtin_rvv_vreinterpret_v_u16m8_u32m8((vuint16m8_t)(op0)) +#define vreinterpret_v_u16mf2_u32mf2(op0) \ +__builtin_rvv_vreinterpret_v_u16mf2_u32mf2((vuint16mf2_t)(op0)) +#define vreinterpret_v_u64m1_u32m1(op0) \ +__builtin_rvv_vreinterpret_v_u64m1_u32m1((vuint64m1_t)(op0)) +#define vreinterpret_v_u64m2_u32m2(op0) \ +__builtin_rvv_vreinterpret_v_u64m2_u32m2((vuint64m2_t)(op0)) +#define vreinterpret_v_u64m4_u32m4(op0) \ +__builtin_rvv_vreinterpret_v_u64m4_u32m4((vuint64m4_t)(op0)) +#define vreinterpret_v_u64m8_u32m8(op0) \ +__builtin_rvv_vreinterpret_v_u64m8_u32m8((vuint64m8_t)(op0)) +#define vreinterpret_v_u8m1_u64m1(op0) \ +__builtin_rvv_vreinterpret_v_u8m1_u64m1((vuint8m1_t)(op0)) +#define vreinterpret_v_u8m2_u64m2(op0) \ +__builtin_rvv_vreinterpret_v_u8m2_u64m2((vuint8m2_t)(op0)) +#define vreinterpret_v_u8m4_u64m4(op0) \ +__builtin_rvv_vreinterpret_v_u8m4_u64m4((vuint8m4_t)(op0)) +#define vreinterpret_v_u8m8_u64m8(op0) \ +__builtin_rvv_vreinterpret_v_u8m8_u64m8((vuint8m8_t)(op0)) +#define vreinterpret_v_u16m1_u64m1(op0) \ +__builtin_rvv_vreinterpret_v_u16m1_u64m1((vuint16m1_t)(op0)) +#define vreinterpret_v_u16m2_u64m2(op0) \ +__builtin_rvv_vreinterpret_v_u16m2_u64m2((vuint16m2_t)(op0)) +#define vreinterpret_v_u16m4_u64m4(op0) \ +__builtin_rvv_vreinterpret_v_u16m4_u64m4((vuint16m4_t)(op0)) +#define vreinterpret_v_u16m8_u64m8(op0) \ +__builtin_rvv_vreinterpret_v_u16m8_u64m8((vuint16m8_t)(op0)) +#define vreinterpret_v_u32m1_u64m1(op0) \ +__builtin_rvv_vreinterpret_v_u32m1_u64m1((vuint32m1_t)(op0)) +#define vreinterpret_v_u32m2_u64m2(op0) \ +__builtin_rvv_vreinterpret_v_u32m2_u64m2((vuint32m2_t)(op0)) +#define vreinterpret_v_u32m4_u64m4(op0) \ +__builtin_rvv_vreinterpret_v_u32m4_u64m4((vuint32m4_t)(op0)) +#define vreinterpret_v_u32m8_u64m8(op0) \ +__builtin_rvv_vreinterpret_v_u32m8_u64m8((vuint32m8_t)(op0)) +#define vreinterpret_v_u16m1_u8m1(op0) \ +__builtin_rvv_vreinterpret_v_u16m1_u8m1((vuint16m1_t)(op0)) +#define vreinterpret_v_u16m2_u8m2(op0) \ +__builtin_rvv_vreinterpret_v_u16m2_u8m2((vuint16m2_t)(op0)) +#define vreinterpret_v_u16m4_u8m4(op0) \ +__builtin_rvv_vreinterpret_v_u16m4_u8m4((vuint16m4_t)(op0)) +#define vreinterpret_v_u16m8_u8m8(op0) \ +__builtin_rvv_vreinterpret_v_u16m8_u8m8((vuint16m8_t)(op0)) +#define vreinterpret_v_u16mf2_u8mf2(op0) \ +__builtin_rvv_vreinterpret_v_u16mf2_u8mf2((vuint16mf2_t)(op0)) +#define vreinterpret_v_u16mf4_u8mf4(op0) \ +__builtin_rvv_vreinterpret_v_u16mf4_u8mf4((vuint16mf4_t)(op0)) +#define vreinterpret_v_u32m1_u8m1(op0) \ +__builtin_rvv_vreinterpret_v_u32m1_u8m1((vuint32m1_t)(op0)) +#define vreinterpret_v_u32m2_u8m2(op0) \ +__builtin_rvv_vreinterpret_v_u32m2_u8m2((vuint32m2_t)(op0)) +#define vreinterpret_v_u32m4_u8m4(op0) \ +__builtin_rvv_vreinterpret_v_u32m4_u8m4((vuint32m4_t)(op0)) +#define vreinterpret_v_u32m8_u8m8(op0) \ +__builtin_rvv_vreinterpret_v_u32m8_u8m8((vuint32m8_t)(op0)) +#define vreinterpret_v_u32mf2_u8mf2(op0) \ +__builtin_rvv_vreinterpret_v_u32mf2_u8mf2((vuint32mf2_t)(op0)) +#define vreinterpret_v_u64m1_u8m1(op0) \ +__builtin_rvv_vreinterpret_v_u64m1_u8m1((vuint64m1_t)(op0)) +#define vreinterpret_v_u64m2_u8m2(op0) \ +__builtin_rvv_vreinterpret_v_u64m2_u8m2((vuint64m2_t)(op0)) +#define vreinterpret_v_u64m4_u8m4(op0) \ +__builtin_rvv_vreinterpret_v_u64m4_u8m4((vuint64m4_t)(op0)) +#define vreinterpret_v_u64m8_u8m8(op0) \ +__builtin_rvv_vreinterpret_v_u64m8_u8m8((vuint64m8_t)(op0)) +#define vreinterpret_v_i8m1_u8m1(op0) \ +__builtin_rvv_vreinterpret_v_i8m1_u8m1((vint8m1_t)(op0)) +#define vreinterpret_v_i8m2_u8m2(op0) \ +__builtin_rvv_vreinterpret_v_i8m2_u8m2((vint8m2_t)(op0)) +#define vreinterpret_v_i8m4_u8m4(op0) \ +__builtin_rvv_vreinterpret_v_i8m4_u8m4((vint8m4_t)(op0)) +#define vreinterpret_v_i8m8_u8m8(op0) \ +__builtin_rvv_vreinterpret_v_i8m8_u8m8((vint8m8_t)(op0)) +#define vreinterpret_v_i8mf2_u8mf2(op0) \ +__builtin_rvv_vreinterpret_v_i8mf2_u8mf2((vint8mf2_t)(op0)) +#define vreinterpret_v_i8mf4_u8mf4(op0) \ +__builtin_rvv_vreinterpret_v_i8mf4_u8mf4((vint8mf4_t)(op0)) +#define vreinterpret_v_i8mf8_u8mf8(op0) \ +__builtin_rvv_vreinterpret_v_i8mf8_u8mf8((vint8mf8_t)(op0)) +#define vreinterpret_v_i16m1_u16m1(op0) \ +__builtin_rvv_vreinterpret_v_i16m1_u16m1((vint16m1_t)(op0)) +#define vreinterpret_v_i16m2_u16m2(op0) \ +__builtin_rvv_vreinterpret_v_i16m2_u16m2((vint16m2_t)(op0)) +#define vreinterpret_v_i16m4_u16m4(op0) \ +__builtin_rvv_vreinterpret_v_i16m4_u16m4((vint16m4_t)(op0)) +#define vreinterpret_v_i16m8_u16m8(op0) \ +__builtin_rvv_vreinterpret_v_i16m8_u16m8((vint16m8_t)(op0)) +#define vreinterpret_v_i16mf2_u16mf2(op0) \ +__builtin_rvv_vreinterpret_v_i16mf2_u16mf2((vint16mf2_t)(op0)) +#define vreinterpret_v_i16mf4_u16mf4(op0) \ +__builtin_rvv_vreinterpret_v_i16mf4_u16mf4((vint16mf4_t)(op0)) +#define vreinterpret_v_i32m1_u32m1(op0) \ +__builtin_rvv_vreinterpret_v_i32m1_u32m1((vint32m1_t)(op0)) +#define vreinterpret_v_i32m2_u32m2(op0) \ +__builtin_rvv_vreinterpret_v_i32m2_u32m2((vint32m2_t)(op0)) +#define vreinterpret_v_i32m4_u32m4(op0) \ +__builtin_rvv_vreinterpret_v_i32m4_u32m4((vint32m4_t)(op0)) +#define vreinterpret_v_i32m8_u32m8(op0) \ +__builtin_rvv_vreinterpret_v_i32m8_u32m8((vint32m8_t)(op0)) +#define vreinterpret_v_i32mf2_u32mf2(op0) \ +__builtin_rvv_vreinterpret_v_i32mf2_u32mf2((vint32mf2_t)(op0)) +#define vreinterpret_v_i64m1_u64m1(op0) \ +__builtin_rvv_vreinterpret_v_i64m1_u64m1((vint64m1_t)(op0)) +#define vreinterpret_v_i64m2_u64m2(op0) \ +__builtin_rvv_vreinterpret_v_i64m2_u64m2((vint64m2_t)(op0)) +#define vreinterpret_v_i64m4_u64m4(op0) \ +__builtin_rvv_vreinterpret_v_i64m4_u64m4((vint64m4_t)(op0)) +#define vreinterpret_v_i64m8_u64m8(op0) \ +__builtin_rvv_vreinterpret_v_i64m8_u64m8((vint64m8_t)(op0)) +#define vse1_v_b8(op0, op1, op2) \ +__builtin_rvv_vse1_v_b8((uint8_t *)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vse1_v_b4(op0, op1, op2) \ +__builtin_rvv_vse1_v_b4((uint8_t *)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vse1_v_b2(op0, op1, op2) \ +__builtin_rvv_vse1_v_b2((uint8_t *)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vse1_v_b1(op0, op1, op2) \ +__builtin_rvv_vse1_v_b1((uint8_t *)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vse1_v_b16(op0, op1, op2) \ +__builtin_rvv_vse1_v_b16((uint8_t *)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vse1_v_b32(op0, op1, op2) \ +__builtin_rvv_vse1_v_b32((uint8_t *)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vse1_v_b64(op0, op1, op2) \ +__builtin_rvv_vse1_v_b64((uint8_t *)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vsext_vf2_i16mf4(op0, op1) \ +__builtin_rvv_vsext_vf2_i16mf4((vint8mf8_t)(op0), (size_t)(op1)) +#define vsext_vf2_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsext_vf2_i16mf2(op0, op1) \ +__builtin_rvv_vsext_vf2_i16mf2((vint8mf4_t)(op0), (size_t)(op1)) +#define vsext_vf2_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsext_vf2_i16m1(op0, op1) \ +__builtin_rvv_vsext_vf2_i16m1((vint8mf2_t)(op0), (size_t)(op1)) +#define vsext_vf2_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsext_vf2_i16m2(op0, op1) \ +__builtin_rvv_vsext_vf2_i16m2((vint8m1_t)(op0), (size_t)(op1)) +#define vsext_vf2_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsext_vf2_i16m4(op0, op1) \ +__builtin_rvv_vsext_vf2_i16m4((vint8m2_t)(op0), (size_t)(op1)) +#define vsext_vf2_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsext_vf2_i16m8(op0, op1) \ +__builtin_rvv_vsext_vf2_i16m8((vint8m4_t)(op0), (size_t)(op1)) +#define vsext_vf2_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsext_vf2_i32mf2(op0, op1) \ +__builtin_rvv_vsext_vf2_i32mf2((vint16mf4_t)(op0), (size_t)(op1)) +#define vsext_vf2_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsext_vf2_i32m1(op0, op1) \ +__builtin_rvv_vsext_vf2_i32m1((vint16mf2_t)(op0), (size_t)(op1)) +#define vsext_vf2_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsext_vf2_i32m2(op0, op1) \ +__builtin_rvv_vsext_vf2_i32m2((vint16m1_t)(op0), (size_t)(op1)) +#define vsext_vf2_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsext_vf2_i32m4(op0, op1) \ +__builtin_rvv_vsext_vf2_i32m4((vint16m2_t)(op0), (size_t)(op1)) +#define vsext_vf2_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsext_vf2_i32m8(op0, op1) \ +__builtin_rvv_vsext_vf2_i32m8((vint16m4_t)(op0), (size_t)(op1)) +#define vsext_vf2_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsext_vf2_i64m1(op0, op1) \ +__builtin_rvv_vsext_vf2_i64m1((vint32mf2_t)(op0), (size_t)(op1)) +#define vsext_vf2_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsext_vf2_i64m2(op0, op1) \ +__builtin_rvv_vsext_vf2_i64m2((vint32m1_t)(op0), (size_t)(op1)) +#define vsext_vf2_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsext_vf2_i64m4(op0, op1) \ +__builtin_rvv_vsext_vf2_i64m4((vint32m2_t)(op0), (size_t)(op1)) +#define vsext_vf2_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsext_vf2_i64m8(op0, op1) \ +__builtin_rvv_vsext_vf2_i64m8((vint32m4_t)(op0), (size_t)(op1)) +#define vsext_vf2_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsext_vf4_i32mf2(op0, op1) \ +__builtin_rvv_vsext_vf4_i32mf2((vint8mf8_t)(op0), (size_t)(op1)) +#define vsext_vf4_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsext_vf4_i32m1(op0, op1) \ +__builtin_rvv_vsext_vf4_i32m1((vint8mf4_t)(op0), (size_t)(op1)) +#define vsext_vf4_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsext_vf4_i32m2(op0, op1) \ +__builtin_rvv_vsext_vf4_i32m2((vint8mf2_t)(op0), (size_t)(op1)) +#define vsext_vf4_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsext_vf4_i32m4(op0, op1) \ +__builtin_rvv_vsext_vf4_i32m4((vint8m1_t)(op0), (size_t)(op1)) +#define vsext_vf4_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsext_vf4_i32m8(op0, op1) \ +__builtin_rvv_vsext_vf4_i32m8((vint8m2_t)(op0), (size_t)(op1)) +#define vsext_vf4_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsext_vf4_i64m1(op0, op1) \ +__builtin_rvv_vsext_vf4_i64m1((vint16mf4_t)(op0), (size_t)(op1)) +#define vsext_vf4_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsext_vf4_i64m2(op0, op1) \ +__builtin_rvv_vsext_vf4_i64m2((vint16mf2_t)(op0), (size_t)(op1)) +#define vsext_vf4_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsext_vf4_i64m4(op0, op1) \ +__builtin_rvv_vsext_vf4_i64m4((vint16m1_t)(op0), (size_t)(op1)) +#define vsext_vf4_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsext_vf4_i64m8(op0, op1) \ +__builtin_rvv_vsext_vf4_i64m8((vint16m2_t)(op0), (size_t)(op1)) +#define vsext_vf4_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsext_vf8_i64m1(op0, op1) \ +__builtin_rvv_vsext_vf8_i64m1((vint8mf8_t)(op0), (size_t)(op1)) +#define vsext_vf8_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf8_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsext_vf8_i64m2(op0, op1) \ +__builtin_rvv_vsext_vf8_i64m2((vint8mf4_t)(op0), (size_t)(op1)) +#define vsext_vf8_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf8_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsext_vf8_i64m4(op0, op1) \ +__builtin_rvv_vsext_vf8_i64m4((vint8mf2_t)(op0), (size_t)(op1)) +#define vsext_vf8_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf8_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsext_vf8_i64m8(op0, op1) \ +__builtin_rvv_vsext_vf8_i64m8((vint8m1_t)(op0), (size_t)(op1)) +#define vsext_vf8_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf8_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vundefined_i8m1() \ +__builtin_rvv_vundefined_i8m1() +#define vundefined_i8m2() \ +__builtin_rvv_vundefined_i8m2() +#define vundefined_i8m4() \ +__builtin_rvv_vundefined_i8m4() +#define vundefined_i8m8() \ +__builtin_rvv_vundefined_i8m8() +#define vundefined_i8mf2() \ +__builtin_rvv_vundefined_i8mf2() +#define vundefined_i8mf4() \ +__builtin_rvv_vundefined_i8mf4() +#define vundefined_i8mf8() \ +__builtin_rvv_vundefined_i8mf8() +#define vundefined_i16m1() \ +__builtin_rvv_vundefined_i16m1() +#define vundefined_i16m2() \ +__builtin_rvv_vundefined_i16m2() +#define vundefined_i16m4() \ +__builtin_rvv_vundefined_i16m4() +#define vundefined_i16m8() \ +__builtin_rvv_vundefined_i16m8() +#define vundefined_i16mf2() \ +__builtin_rvv_vundefined_i16mf2() +#define vundefined_i16mf4() \ +__builtin_rvv_vundefined_i16mf4() +#define vundefined_i32m1() \ +__builtin_rvv_vundefined_i32m1() +#define vundefined_i32m2() \ +__builtin_rvv_vundefined_i32m2() +#define vundefined_i32m4() \ +__builtin_rvv_vundefined_i32m4() +#define vundefined_i32m8() \ +__builtin_rvv_vundefined_i32m8() +#define vundefined_i32mf2() \ +__builtin_rvv_vundefined_i32mf2() +#define vundefined_i64m1() \ +__builtin_rvv_vundefined_i64m1() +#define vundefined_i64m2() \ +__builtin_rvv_vundefined_i64m2() +#define vundefined_i64m4() \ +__builtin_rvv_vundefined_i64m4() +#define vundefined_i64m8() \ +__builtin_rvv_vundefined_i64m8() +#define vundefined_u8m1() \ +__builtin_rvv_vundefined_u8m1() +#define vundefined_u8m2() \ +__builtin_rvv_vundefined_u8m2() +#define vundefined_u8m4() \ +__builtin_rvv_vundefined_u8m4() +#define vundefined_u8m8() \ +__builtin_rvv_vundefined_u8m8() +#define vundefined_u8mf2() \ +__builtin_rvv_vundefined_u8mf2() +#define vundefined_u8mf4() \ +__builtin_rvv_vundefined_u8mf4() +#define vundefined_u8mf8() \ +__builtin_rvv_vundefined_u8mf8() +#define vundefined_u16m1() \ +__builtin_rvv_vundefined_u16m1() +#define vundefined_u16m2() \ +__builtin_rvv_vundefined_u16m2() +#define vundefined_u16m4() \ +__builtin_rvv_vundefined_u16m4() +#define vundefined_u16m8() \ +__builtin_rvv_vundefined_u16m8() +#define vundefined_u16mf2() \ +__builtin_rvv_vundefined_u16mf2() +#define vundefined_u16mf4() \ +__builtin_rvv_vundefined_u16mf4() +#define vundefined_u32m1() \ +__builtin_rvv_vundefined_u32m1() +#define vundefined_u32m2() \ +__builtin_rvv_vundefined_u32m2() +#define vundefined_u32m4() \ +__builtin_rvv_vundefined_u32m4() +#define vundefined_u32m8() \ +__builtin_rvv_vundefined_u32m8() +#define vundefined_u32mf2() \ +__builtin_rvv_vundefined_u32mf2() +#define vundefined_u64m1() \ +__builtin_rvv_vundefined_u64m1() +#define vundefined_u64m2() \ +__builtin_rvv_vundefined_u64m2() +#define vundefined_u64m4() \ +__builtin_rvv_vundefined_u64m4() +#define vundefined_u64m8() \ +__builtin_rvv_vundefined_u64m8() +#define vzext_vf2_u16mf4(op0, op1) \ +__builtin_rvv_vzext_vf2_u16mf4((vuint8mf8_t)(op0), (size_t)(op1)) +#define vzext_vf2_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vzext_vf2_u16mf2(op0, op1) \ +__builtin_rvv_vzext_vf2_u16mf2((vuint8mf4_t)(op0), (size_t)(op1)) +#define vzext_vf2_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vzext_vf2_u16m1(op0, op1) \ +__builtin_rvv_vzext_vf2_u16m1((vuint8mf2_t)(op0), (size_t)(op1)) +#define vzext_vf2_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vzext_vf2_u16m2(op0, op1) \ +__builtin_rvv_vzext_vf2_u16m2((vuint8m1_t)(op0), (size_t)(op1)) +#define vzext_vf2_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vzext_vf2_u16m4(op0, op1) \ +__builtin_rvv_vzext_vf2_u16m4((vuint8m2_t)(op0), (size_t)(op1)) +#define vzext_vf2_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vzext_vf2_u16m8(op0, op1) \ +__builtin_rvv_vzext_vf2_u16m8((vuint8m4_t)(op0), (size_t)(op1)) +#define vzext_vf2_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vzext_vf2_u32mf2(op0, op1) \ +__builtin_rvv_vzext_vf2_u32mf2((vuint16mf4_t)(op0), (size_t)(op1)) +#define vzext_vf2_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vzext_vf2_u32m1(op0, op1) \ +__builtin_rvv_vzext_vf2_u32m1((vuint16mf2_t)(op0), (size_t)(op1)) +#define vzext_vf2_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vzext_vf2_u32m2(op0, op1) \ +__builtin_rvv_vzext_vf2_u32m2((vuint16m1_t)(op0), (size_t)(op1)) +#define vzext_vf2_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vzext_vf2_u32m4(op0, op1) \ +__builtin_rvv_vzext_vf2_u32m4((vuint16m2_t)(op0), (size_t)(op1)) +#define vzext_vf2_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vzext_vf2_u32m8(op0, op1) \ +__builtin_rvv_vzext_vf2_u32m8((vuint16m4_t)(op0), (size_t)(op1)) +#define vzext_vf2_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vzext_vf2_u64m1(op0, op1) \ +__builtin_rvv_vzext_vf2_u64m1((vuint32mf2_t)(op0), (size_t)(op1)) +#define vzext_vf2_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vzext_vf2_u64m2(op0, op1) \ +__builtin_rvv_vzext_vf2_u64m2((vuint32m1_t)(op0), (size_t)(op1)) +#define vzext_vf2_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vzext_vf2_u64m4(op0, op1) \ +__builtin_rvv_vzext_vf2_u64m4((vuint32m2_t)(op0), (size_t)(op1)) +#define vzext_vf2_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vzext_vf2_u64m8(op0, op1) \ +__builtin_rvv_vzext_vf2_u64m8((vuint32m4_t)(op0), (size_t)(op1)) +#define vzext_vf2_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vzext_vf4_u32mf2(op0, op1) \ +__builtin_rvv_vzext_vf4_u32mf2((vuint8mf8_t)(op0), (size_t)(op1)) +#define vzext_vf4_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vzext_vf4_u32m1(op0, op1) \ +__builtin_rvv_vzext_vf4_u32m1((vuint8mf4_t)(op0), (size_t)(op1)) +#define vzext_vf4_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vzext_vf4_u32m2(op0, op1) \ +__builtin_rvv_vzext_vf4_u32m2((vuint8mf2_t)(op0), (size_t)(op1)) +#define vzext_vf4_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vzext_vf4_u32m4(op0, op1) \ +__builtin_rvv_vzext_vf4_u32m4((vuint8m1_t)(op0), (size_t)(op1)) +#define vzext_vf4_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vzext_vf4_u32m8(op0, op1) \ +__builtin_rvv_vzext_vf4_u32m8((vuint8m2_t)(op0), (size_t)(op1)) +#define vzext_vf4_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vzext_vf4_u64m1(op0, op1) \ +__builtin_rvv_vzext_vf4_u64m1((vuint16mf4_t)(op0), (size_t)(op1)) +#define vzext_vf4_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vzext_vf4_u64m2(op0, op1) \ +__builtin_rvv_vzext_vf4_u64m2((vuint16mf2_t)(op0), (size_t)(op1)) +#define vzext_vf4_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vzext_vf4_u64m4(op0, op1) \ +__builtin_rvv_vzext_vf4_u64m4((vuint16m1_t)(op0), (size_t)(op1)) +#define vzext_vf4_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vzext_vf4_u64m8(op0, op1) \ +__builtin_rvv_vzext_vf4_u64m8((vuint16m2_t)(op0), (size_t)(op1)) +#define vzext_vf4_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vzext_vf8_u64m1(op0, op1) \ +__builtin_rvv_vzext_vf8_u64m1((vuint8mf8_t)(op0), (size_t)(op1)) +#define vzext_vf8_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf8_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vzext_vf8_u64m2(op0, op1) \ +__builtin_rvv_vzext_vf8_u64m2((vuint8mf4_t)(op0), (size_t)(op1)) +#define vzext_vf8_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf8_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vzext_vf8_u64m4(op0, op1) \ +__builtin_rvv_vzext_vf8_u64m4((vuint8mf2_t)(op0), (size_t)(op1)) +#define vzext_vf8_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf8_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vzext_vf8_u64m8(op0, op1) \ +__builtin_rvv_vzext_vf8_u64m8((vuint8m1_t)(op0), (size_t)(op1)) +#define vzext_vf8_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf8_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#if defined(__riscv_f) +#define vsse32_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsse32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsse32_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_f32m2((float *)(op0), (ptrdiff_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsse32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsse32_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_f32m4((float *)(op0), (ptrdiff_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsse32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsse32_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_f32m8((float *)(op0), (ptrdiff_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsse32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsse32_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsse32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f32m1((const float *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f32m2((const float *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f32m4((const float *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f32m8((const float *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f32mf2((const float *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei16_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f32m1((const float *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f32m2((const float *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f32m4((const float *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f32m8((const float *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f32mf2((const float *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei32_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f32m1((const float *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f32m2((const float *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f32m4((const float *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f32m8((const float *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f32mf2((const float *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei64_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f32m1((const float *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f32m2((const float *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f32m4((const float *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f32mf2((const float *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f32m1((const float *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f32m2((const float *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f32m4((const float *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f32m8((const float *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f32mf2((const float *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei16_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f32m1((const float *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f32m2((const float *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f32m4((const float *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f32m8((const float *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f32mf2((const float *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei32_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f32m1((const float *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f32m2((const float *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f32m4((const float *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f32m8((const float *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f32mf2((const float *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei64_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f32m1((const float *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f32m2((const float *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f32m4((const float *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f32mf2((const float *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmv_v_v_f32m1(op0, op1) \ +__builtin_rvv_vmv_v_v_f32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_f32m2(op0, op1) \ +__builtin_rvv_vmv_v_v_f32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_f32m4(op0, op1) \ +__builtin_rvv_vmv_v_v_f32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_f32m8(op0, op1) \ +__builtin_rvv_vmv_v_v_f32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_f32mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfadd_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfadd_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfadd_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfadd_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfadd_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfadd_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfadd_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfadd_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfadd_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfadd_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfadd_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfadd_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfadd_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfadd_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfadd_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfadd_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfadd_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfadd_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfadd_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfadd_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsub_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfsub_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfsub_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfsub_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfsub_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfsub_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfsub_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfsub_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfsub_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfsub_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfsub_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsub_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsub_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsub_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsub_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsub_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsub_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsub_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsub_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsub_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrsub_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrsub_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrsub_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrsub_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrsub_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrsub_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrsub_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrsub_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrsub_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrsub_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmul_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfmul_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmul_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfmul_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmul_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfmul_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmul_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfmul_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmul_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfmul_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmul_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmul_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmul_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmul_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmul_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmul_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmul_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmul_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmul_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmul_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfdiv_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfdiv_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfdiv_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfdiv_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfdiv_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfdiv_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfdiv_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfdiv_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfdiv_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfdiv_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfdiv_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrdiv_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrdiv_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrdiv_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrdiv_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrdiv_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrdiv_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrdiv_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrdiv_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrdiv_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrdiv_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmacc_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmin_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfmin_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmin_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfmin_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmin_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfmin_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmin_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfmin_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmin_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfmin_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmin_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmin_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmin_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmin_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmin_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmin_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmin_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmin_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmin_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmin_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmax_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfmax_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmax_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfmax_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmax_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfmax_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmax_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfmax_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmax_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfmax_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmax_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmax_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmax_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmax_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmax_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmax_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmax_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmax_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmax_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmax_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnj_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfsgnj_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnj_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnj_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnj_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnj_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnj_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnj_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnj_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnj_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnj_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfabs_v_f32m1(op0, op1) \ +__builtin_rvv_vfabs_v_f32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfabs_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfabs_v_f32m2(op0, op1) \ +__builtin_rvv_vfabs_v_f32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfabs_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfabs_v_f32m4(op0, op1) \ +__builtin_rvv_vfabs_v_f32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfabs_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfabs_v_f32m8(op0, op1) \ +__builtin_rvv_vfabs_v_f32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfabs_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfabs_v_f32mf2(op0, op1) \ +__builtin_rvv_vfabs_v_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfabs_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vmfeq_vv_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f32m1_b32((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f32m2_b16((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f32m4_b8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f32m8_b4((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f32mf2_b64((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vmfeq_vf_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f32m1_b32((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfeq_vf_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfeq_vf_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f32m2_b16((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfeq_vf_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfeq_vf_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f32m4_b8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfeq_vf_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfeq_vf_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f32m8_b4((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfeq_vf_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfeq_vf_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f32mf2_b64((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfeq_vf_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfne_vv_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f32m1_b32((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vmfne_vv_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vmfne_vv_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f32m2_b16((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vmfne_vv_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vmfne_vv_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f32m4_b8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vmfne_vv_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vmfne_vv_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f32m8_b4((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vmfne_vv_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vmfne_vv_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f32mf2_b64((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vmfne_vv_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vmfne_vf_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f32m1_b32((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfne_vf_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfne_vf_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f32m2_b16((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfne_vf_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfne_vf_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f32m4_b8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfne_vf_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfne_vf_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f32m8_b4((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfne_vf_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfne_vf_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f32mf2_b64((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfne_vf_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmflt_vv_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f32m1_b32((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vmflt_vv_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vmflt_vv_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f32m2_b16((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vmflt_vv_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vmflt_vv_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f32m4_b8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vmflt_vv_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vmflt_vv_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f32m8_b4((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vmflt_vv_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vmflt_vv_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f32mf2_b64((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vmflt_vv_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vmflt_vf_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f32m1_b32((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vmflt_vf_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vmflt_vf_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f32m2_b16((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmflt_vf_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmflt_vf_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f32m4_b8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vmflt_vf_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vmflt_vf_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f32m8_b4((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vmflt_vf_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vmflt_vf_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f32mf2_b64((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmflt_vf_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfle_vv_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f32m1_b32((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vmfle_vv_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vmfle_vv_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f32m2_b16((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vmfle_vv_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vmfle_vv_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f32m4_b8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vmfle_vv_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vmfle_vv_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f32m8_b4((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vmfle_vv_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vmfle_vv_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f32mf2_b64((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vmfle_vv_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vmfle_vf_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f32m1_b32((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfle_vf_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfle_vf_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f32m2_b16((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfle_vf_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfle_vf_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f32m4_b8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfle_vf_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfle_vf_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f32m8_b4((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfle_vf_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfle_vf_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f32mf2_b64((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfle_vf_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfgt_vv_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f32m1_b32((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f32m2_b16((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f32m4_b8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f32m8_b4((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f32mf2_b64((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vmfgt_vf_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f32m1_b32((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfgt_vf_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfgt_vf_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f32m2_b16((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfgt_vf_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfgt_vf_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f32m4_b8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfgt_vf_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfgt_vf_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f32m8_b4((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfgt_vf_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfgt_vf_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f32mf2_b64((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfgt_vf_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfge_vv_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f32m1_b32((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vmfge_vv_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vmfge_vv_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f32m2_b16((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vmfge_vv_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vmfge_vv_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f32m4_b8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vmfge_vv_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vmfge_vv_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f32m8_b4((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vmfge_vv_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vmfge_vv_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f32mf2_b64((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vmfge_vv_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vmfge_vf_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f32m1_b32((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfge_vf_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfge_vf_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f32m2_b16((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfge_vf_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfge_vf_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f32m4_b8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfge_vf_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfge_vf_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f32m8_b4((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfge_vf_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfge_vf_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f32mf2_b64((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfge_vf_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmerge_vvm_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f32m1((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f32m2((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f32m4((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f32m8((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f32mf2((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmerge_vfm_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f32m1((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (size_t)(op3)) +#define vfmerge_vfm_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f32m2((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (size_t)(op3)) +#define vfmerge_vfm_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f32m4((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (size_t)(op3)) +#define vfmerge_vfm_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f32m8((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (size_t)(op3)) +#define vfmerge_vfm_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f32mf2((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (size_t)(op3)) +#define vfmv_v_f_f32m1(op0, op1) \ +__builtin_rvv_vfmv_v_f_f32m1((float)(op0), (size_t)(op1)) +#define vfmv_v_f_f32m2(op0, op1) \ +__builtin_rvv_vfmv_v_f_f32m2((float)(op0), (size_t)(op1)) +#define vfmv_v_f_f32m4(op0, op1) \ +__builtin_rvv_vfmv_v_f_f32m4((float)(op0), (size_t)(op1)) +#define vfmv_v_f_f32m8(op0, op1) \ +__builtin_rvv_vfmv_v_f_f32m8((float)(op0), (size_t)(op1)) +#define vfmv_v_f_f32mf2(op0, op1) \ +__builtin_rvv_vfmv_v_f_f32mf2((float)(op0), (size_t)(op1)) +#define vfredmax_vs_f32m1_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f32m1_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f32m1_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f32m1_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f32m2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f32m2_f32m1((vfloat32m1_t)(op0), (vfloat32m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f32m2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f32m2_f32m1_m((vbool16_t)(op0), (vfloat32m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f32m4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f32m4_f32m1((vfloat32m1_t)(op0), (vfloat32m4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f32m4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f32m4_f32m1_m((vbool8_t)(op0), (vfloat32m1_t)(op1), (vfloat32m4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f32m8_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f32m8_f32m1((vfloat32m1_t)(op0), (vfloat32m8_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f32m8_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f32m8_f32m1_m((vbool4_t)(op0), (vfloat32m1_t)(op1), (vfloat32m8_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f32mf2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f32mf2_f32m1((vfloat32m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f32mf2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f32mf2_f32m1_m((vbool64_t)(op0), (vfloat32m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f32m1_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f32m1_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f32m1_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f32m1_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f32m2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f32m2_f32m1((vfloat32m1_t)(op0), (vfloat32m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f32m2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f32m2_f32m1_m((vbool16_t)(op0), (vfloat32m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f32m4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f32m4_f32m1((vfloat32m1_t)(op0), (vfloat32m4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f32m4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f32m4_f32m1_m((vbool8_t)(op0), (vfloat32m1_t)(op1), (vfloat32m4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f32m8_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f32m8_f32m1((vfloat32m1_t)(op0), (vfloat32m8_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f32m8_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f32m8_f32m1_m((vbool4_t)(op0), (vfloat32m1_t)(op1), (vfloat32m8_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f32mf2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f32mf2_f32m1((vfloat32m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f32mf2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f32mf2_f32m1_m((vbool64_t)(op0), (vfloat32m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f32m1_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f32m1_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f32m1_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f32m1_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f32m2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f32m2_f32m1((vfloat32m1_t)(op0), (vfloat32m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f32m2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f32m2_f32m1_m((vbool16_t)(op0), (vfloat32m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f32m4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f32m4_f32m1((vfloat32m1_t)(op0), (vfloat32m4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f32m4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f32m4_f32m1_m((vbool8_t)(op0), (vfloat32m1_t)(op1), (vfloat32m4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f32m8_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f32m8_f32m1((vfloat32m1_t)(op0), (vfloat32m8_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f32m8_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f32m8_f32m1_m((vbool4_t)(op0), (vfloat32m1_t)(op1), (vfloat32m8_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f32mf2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f32mf2_f32m1((vfloat32m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f32mf2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f32mf2_f32m1_m((vbool64_t)(op0), (vfloat32m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f32m1_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f32m1_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f32m1_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f32m1_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f32m2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f32m2_f32m1((vfloat32m1_t)(op0), (vfloat32m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f32m2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f32m2_f32m1_m((vbool16_t)(op0), (vfloat32m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f32m4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f32m4_f32m1((vfloat32m1_t)(op0), (vfloat32m4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f32m4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f32m4_f32m1_m((vbool8_t)(op0), (vfloat32m1_t)(op1), (vfloat32m4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f32m8_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f32m8_f32m1((vfloat32m1_t)(op0), (vfloat32m8_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f32m8_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f32m8_f32m1_m((vbool4_t)(op0), (vfloat32m1_t)(op1), (vfloat32m8_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f32mf2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f32mf2_f32m1((vfloat32m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f32mf2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f32mf2_f32m1_m((vbool64_t)(op0), (vfloat32m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmv_f_s_f32m1_f32(op0) \ +__builtin_rvv_vfmv_f_s_f32m1_f32((vfloat32m1_t)(op0)) +#define vfmv_f_s_f32m2_f32(op0) \ +__builtin_rvv_vfmv_f_s_f32m2_f32((vfloat32m2_t)(op0)) +#define vfmv_f_s_f32m4_f32(op0) \ +__builtin_rvv_vfmv_f_s_f32m4_f32((vfloat32m4_t)(op0)) +#define vfmv_f_s_f32m8_f32(op0) \ +__builtin_rvv_vfmv_f_s_f32m8_f32((vfloat32m8_t)(op0)) +#define vfmv_f_s_f32mf2_f32(op0) \ +__builtin_rvv_vfmv_f_s_f32mf2_f32((vfloat32mf2_t)(op0)) +#define vfmv_s_f_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmv_s_f_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmv_s_f_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmv_s_f_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmv_s_f_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vslideup_vx_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vfslide1up_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1up_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1up_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1up_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1up_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1up_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1up_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1up_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1up_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1up_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1down_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1down_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1down_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1down_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1down_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1down_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1down_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1down_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1down_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1down_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vrgather_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f32m1((vfloat32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f32m2((vfloat32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f32m4((vfloat32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f32m8((vfloat32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f32mf2((vfloat32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vrgather_vx_f32m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f32m1((vfloat32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f32m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f32m2((vfloat32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f32m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f32m4((vfloat32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f32m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f32m8((vfloat32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f32mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f32m1((vfloat32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f32m2((vfloat32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f32m4((vfloat32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f32m8((vfloat32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f32mf2((vfloat32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vcompress_vm_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f32m1((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f32m2((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f32m4((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f32m8((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f32mf2((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vget_v_f32m2_f32m1(op0, op1) \ +__builtin_rvv_vget_v_f32m2_f32m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vget_v_f32m4_f32m1(op0, op1) \ +__builtin_rvv_vget_v_f32m4_f32m1((vfloat32m4_t)(op0), (size_t)(op1)) +#define vget_v_f32m8_f32m1(op0, op1) \ +__builtin_rvv_vget_v_f32m8_f32m1((vfloat32m8_t)(op0), (size_t)(op1)) +#define vget_v_f32m4_f32m2(op0, op1) \ +__builtin_rvv_vget_v_f32m4_f32m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vget_v_f32m8_f32m2(op0, op1) \ +__builtin_rvv_vget_v_f32m8_f32m2((vfloat32m8_t)(op0), (size_t)(op1)) +#define vget_v_f32m8_f32m4(op0, op1) \ +__builtin_rvv_vget_v_f32m8_f32m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vset_v_f32m1_f32m2(op0, op1, op2) \ +__builtin_rvv_vset_v_f32m1_f32m2((vfloat32m2_t)(op0), (size_t)(op1), (vfloat32m1_t)(op2)) +#define vset_v_f32m1_f32m4(op0, op1, op2) \ +__builtin_rvv_vset_v_f32m1_f32m4((vfloat32m4_t)(op0), (size_t)(op1), (vfloat32m1_t)(op2)) +#define vset_v_f32m2_f32m4(op0, op1, op2) \ +__builtin_rvv_vset_v_f32m2_f32m4((vfloat32m4_t)(op0), (size_t)(op1), (vfloat32m2_t)(op2)) +#define vsuxei8_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f32m4((float *)(op0), (vuint8m1_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint8m1_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f32m8((float *)(op0), (vuint8m2_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint8m2_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vset_v_f32m1_f32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_f32m1_f32m8((vfloat32m8_t)(op0), (size_t)(op1), (vfloat32m1_t)(op2)) +#define vset_v_f32m2_f32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_f32m2_f32m8((vfloat32m8_t)(op0), (size_t)(op1), (vfloat32m2_t)(op2)) +#define vset_v_f32m4_f32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_f32m4_f32m8((vfloat32m8_t)(op0), (size_t)(op1), (vfloat32m4_t)(op2)) +#define vsuxei16_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f32m4((float *)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint16m2_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f32m8((float *)(op0), (vuint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint16m4_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f32m4((float *)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f32m8((float *)(op0), (vuint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f32m4((float *)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint64m8_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f32m4((float *)(op0), (vuint8m1_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint8m1_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f32m8((float *)(op0), (vuint8m2_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint8m2_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f32m4((float *)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint16m2_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f32m8((float *)(op0), (vuint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint16m4_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f32m4((float *)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f32m8((float *)(op0), (vuint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f32m4((float *)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint64m8_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vle32ff_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_f32m1((const float *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_f32m2((const float *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_f32m4((const float *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_f32m8((const float *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_f32mf2((const float *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vfneg_v_f32m1(op0, op1) \ +__builtin_rvv_vfneg_v_f32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfneg_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfneg_v_f32m2(op0, op1) \ +__builtin_rvv_vfneg_v_f32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfneg_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfneg_v_f32m4(op0, op1) \ +__builtin_rvv_vfneg_v_f32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfneg_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfneg_v_f32m8(op0, op1) \ +__builtin_rvv_vfneg_v_f32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfneg_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfneg_v_f32mf2(op0, op1) \ +__builtin_rvv_vfneg_v_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfneg_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vle32_v_f32m1(op0, op1) \ +__builtin_rvv_vle32_v_f32m1((const float *)(op0), (size_t)(op1)) +#define vle32_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (size_t)(op3)) +#define vle32_v_f32m2(op0, op1) \ +__builtin_rvv_vle32_v_f32m2((const float *)(op0), (size_t)(op1)) +#define vle32_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (size_t)(op3)) +#define vle32_v_f32m4(op0, op1) \ +__builtin_rvv_vle32_v_f32m4((const float *)(op0), (size_t)(op1)) +#define vle32_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (size_t)(op3)) +#define vle32_v_f32m8(op0, op1) \ +__builtin_rvv_vle32_v_f32m8((const float *)(op0), (size_t)(op1)) +#define vle32_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (size_t)(op3)) +#define vle32_v_f32mf2(op0, op1) \ +__builtin_rvv_vle32_v_f32mf2((const float *)(op0), (size_t)(op1)) +#define vle32_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (size_t)(op3)) +#define vse32_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vse32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vse32_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vse32_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vse32_v_f32m2((float *)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vse32_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vse32_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vse32_v_f32m4((float *)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vse32_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vse32_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vse32_v_f32m8((float *)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vse32_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vse32_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vse32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vse32_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vlse32_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vlse32_v_f32m1((const float *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vlse32_v_f32m2((const float *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vlse32_v_f32m4((const float *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vlse32_v_f32m8((const float *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vlse32_v_f32mf2((const float *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vfclass_v_u32m1(op0, op1) \ +__builtin_rvv_vfclass_v_u32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfclass_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfclass_v_u32m2(op0, op1) \ +__builtin_rvv_vfclass_v_u32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfclass_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfclass_v_u32m4(op0, op1) \ +__builtin_rvv_vfclass_v_u32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfclass_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfclass_v_u32m8(op0, op1) \ +__builtin_rvv_vfclass_v_u32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfclass_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfclass_v_u32mf2(op0, op1) \ +__builtin_rvv_vfclass_v_u32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfclass_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f32m1(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f32m1((vint32m1_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f32m2(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f32m2((vint32m2_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f32m4(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f32m4((vint32m4_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f32m8(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f32m8((vint32m8_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f32mf2(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f32mf2((vint32mf2_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f32m1(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f32m1((vuint32m1_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f32m2(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f32m2((vuint32m2_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f32m4(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f32m4((vuint32m4_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f32m8(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f32m8((vuint32m8_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f32mf2(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f32mf2((vuint32mf2_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i32m1(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i32m2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i32m4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i32m8(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i32mf2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u32m1(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u32m2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u32m4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u32m8(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u32mf2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i32m1(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i32m2(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i32m4(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i32m8(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i32mf2(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u32m1(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u32m2(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u32m4(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u32m8(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u32mf2(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f32mf2(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f32mf2((vint64m1_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f32m1(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f32m1((vint64m2_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f32m2(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f32m2((vint64m4_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f32m4(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f32m4((vint64m8_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f32mf2(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f32mf2((vuint64m1_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f32m1(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f32m1((vuint64m2_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f32m2(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f32m2((vuint64m4_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f32m4(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f32m4((vuint64m8_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i16mf4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16mf4((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i16mf2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16mf2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i16m1(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i16m2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i16m4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u16mf4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf4((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u16mf2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u16m1(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u16m2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u16m4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i16mf4(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i16mf4((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i16mf2(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i16mf2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i16m1(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i16m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i16m2(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i16m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i16m4(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i16m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u16mf4(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u16mf4((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u16mf2(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u16mf2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u16m1(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u16m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u16m2(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u16m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u16m4(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u16m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfrec7_v_f32m1(op0, op1) \ +__builtin_rvv_vfrec7_v_f32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfrec7_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfrec7_v_f32m2(op0, op1) \ +__builtin_rvv_vfrec7_v_f32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfrec7_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfrec7_v_f32m4(op0, op1) \ +__builtin_rvv_vfrec7_v_f32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfrec7_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfrec7_v_f32m8(op0, op1) \ +__builtin_rvv_vfrec7_v_f32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfrec7_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfrec7_v_f32mf2(op0, op1) \ +__builtin_rvv_vfrec7_v_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfrec7_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f32m1(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f32m2(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f32m4(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f32m8(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f32mf2(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f32m1(op0, op1) \ +__builtin_rvv_vfsqrt_v_f32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f32m2(op0, op1) \ +__builtin_rvv_vfsqrt_v_f32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f32m4(op0, op1) \ +__builtin_rvv_vfsqrt_v_f32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f32m8(op0, op1) \ +__builtin_rvv_vfsqrt_v_f32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f32mf2(op0, op1) \ +__builtin_rvv_vfsqrt_v_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f32mf2((vint16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f32m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f32m1((vint16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f32m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f32m2((vint16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f32m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f32m4((vint16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f32m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f32m8((vint16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f32mf2((vuint16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f32m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m1((vuint16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f32m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m2((vuint16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f32m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m4((vuint16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f32m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m8((vuint16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i64m1(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m1((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i64m2(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i64m4(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m4((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i64m8(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m8((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u64m1(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m1((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u64m2(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u64m4(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m4((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u64m8(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m8((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i64m1(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i64m1((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i64m2(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i64m2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i64m4(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i64m4((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i64m8(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i64m8((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u64m1(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m1((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u64m2(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u64m4(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m4((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u64m8(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m8((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vlmul_ext_v_f32mf2_f32m1(op0) \ +__builtin_rvv_vlmul_ext_v_f32mf2_f32m1((vfloat32mf2_t)(op0)) +#define vlmul_ext_v_f32m1_f32m2(op0) \ +__builtin_rvv_vlmul_ext_v_f32m1_f32m2((vfloat32m1_t)(op0)) +#define vlmul_ext_v_f32mf2_f32m2(op0) \ +__builtin_rvv_vlmul_ext_v_f32mf2_f32m2((vfloat32mf2_t)(op0)) +#define vlmul_ext_v_f32m1_f32m4(op0) \ +__builtin_rvv_vlmul_ext_v_f32m1_f32m4((vfloat32m1_t)(op0)) +#define vlmul_ext_v_f32m2_f32m4(op0) \ +__builtin_rvv_vlmul_ext_v_f32m2_f32m4((vfloat32m2_t)(op0)) +#define vlmul_ext_v_f32mf2_f32m4(op0) \ +__builtin_rvv_vlmul_ext_v_f32mf2_f32m4((vfloat32mf2_t)(op0)) +#define vlmul_ext_v_f32m1_f32m8(op0) \ +__builtin_rvv_vlmul_ext_v_f32m1_f32m8((vfloat32m1_t)(op0)) +#define vlmul_ext_v_f32m2_f32m8(op0) \ +__builtin_rvv_vlmul_ext_v_f32m2_f32m8((vfloat32m2_t)(op0)) +#define vlmul_ext_v_f32m4_f32m8(op0) \ +__builtin_rvv_vlmul_ext_v_f32m4_f32m8((vfloat32m4_t)(op0)) +#define vlmul_ext_v_f32mf2_f32m8(op0) \ +__builtin_rvv_vlmul_ext_v_f32mf2_f32m8((vfloat32mf2_t)(op0)) +#define vlmul_trunc_v_f32m1_f32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m1_f32mf2((vfloat32m1_t)(op0)) +#define vlmul_trunc_v_f32m2_f32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m2_f32mf2((vfloat32m2_t)(op0)) +#define vlmul_trunc_v_f32m4_f32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m4_f32mf2((vfloat32m4_t)(op0)) +#define vlmul_trunc_v_f32m8_f32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m8_f32mf2((vfloat32m8_t)(op0)) +#define vlmul_trunc_v_f32m2_f32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m2_f32m1((vfloat32m2_t)(op0)) +#define vlmul_trunc_v_f32m4_f32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m4_f32m1((vfloat32m4_t)(op0)) +#define vlmul_trunc_v_f32m8_f32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m8_f32m1((vfloat32m8_t)(op0)) +#define vlmul_trunc_v_f32m4_f32m2(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m4_f32m2((vfloat32m4_t)(op0)) +#define vlmul_trunc_v_f32m8_f32m2(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m8_f32m2((vfloat32m8_t)(op0)) +#define vlmul_trunc_v_f32m8_f32m4(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m8_f32m4((vfloat32m8_t)(op0)) +#define vreinterpret_v_i32m1_f32m1(op0) \ +__builtin_rvv_vreinterpret_v_i32m1_f32m1((vint32m1_t)(op0)) +#define vreinterpret_v_i32m2_f32m2(op0) \ +__builtin_rvv_vreinterpret_v_i32m2_f32m2((vint32m2_t)(op0)) +#define vreinterpret_v_i32m4_f32m4(op0) \ +__builtin_rvv_vreinterpret_v_i32m4_f32m4((vint32m4_t)(op0)) +#define vreinterpret_v_i32m8_f32m8(op0) \ +__builtin_rvv_vreinterpret_v_i32m8_f32m8((vint32m8_t)(op0)) +#define vreinterpret_v_i32mf2_f32mf2(op0) \ +__builtin_rvv_vreinterpret_v_i32mf2_f32mf2((vint32mf2_t)(op0)) +#define vreinterpret_v_u32m1_f32m1(op0) \ +__builtin_rvv_vreinterpret_v_u32m1_f32m1((vuint32m1_t)(op0)) +#define vreinterpret_v_u32m2_f32m2(op0) \ +__builtin_rvv_vreinterpret_v_u32m2_f32m2((vuint32m2_t)(op0)) +#define vreinterpret_v_u32m4_f32m4(op0) \ +__builtin_rvv_vreinterpret_v_u32m4_f32m4((vuint32m4_t)(op0)) +#define vreinterpret_v_u32m8_f32m8(op0) \ +__builtin_rvv_vreinterpret_v_u32m8_f32m8((vuint32m8_t)(op0)) +#define vreinterpret_v_u32mf2_f32mf2(op0) \ +__builtin_rvv_vreinterpret_v_u32mf2_f32mf2((vuint32mf2_t)(op0)) +#define vreinterpret_v_f32m1_i32m1(op0) \ +__builtin_rvv_vreinterpret_v_f32m1_i32m1((vfloat32m1_t)(op0)) +#define vreinterpret_v_f32m2_i32m2(op0) \ +__builtin_rvv_vreinterpret_v_f32m2_i32m2((vfloat32m2_t)(op0)) +#define vreinterpret_v_f32m4_i32m4(op0) \ +__builtin_rvv_vreinterpret_v_f32m4_i32m4((vfloat32m4_t)(op0)) +#define vreinterpret_v_f32m8_i32m8(op0) \ +__builtin_rvv_vreinterpret_v_f32m8_i32m8((vfloat32m8_t)(op0)) +#define vreinterpret_v_f32mf2_i32mf2(op0) \ +__builtin_rvv_vreinterpret_v_f32mf2_i32mf2((vfloat32mf2_t)(op0)) +#define vreinterpret_v_f32m1_u32m1(op0) \ +__builtin_rvv_vreinterpret_v_f32m1_u32m1((vfloat32m1_t)(op0)) +#define vreinterpret_v_f32m2_u32m2(op0) \ +__builtin_rvv_vreinterpret_v_f32m2_u32m2((vfloat32m2_t)(op0)) +#define vreinterpret_v_f32m4_u32m4(op0) \ +__builtin_rvv_vreinterpret_v_f32m4_u32m4((vfloat32m4_t)(op0)) +#define vreinterpret_v_f32m8_u32m8(op0) \ +__builtin_rvv_vreinterpret_v_f32m8_u32m8((vfloat32m8_t)(op0)) +#define vreinterpret_v_f32mf2_u32mf2(op0) \ +__builtin_rvv_vreinterpret_v_f32mf2_u32mf2((vfloat32mf2_t)(op0)) +#define vundefined_f32m1() \ +__builtin_rvv_vundefined_f32m1() +#define vundefined_f32m2() \ +__builtin_rvv_vundefined_f32m2() +#define vundefined_f32m4() \ +__builtin_rvv_vundefined_f32m4() +#define vundefined_f32m8() \ +__builtin_rvv_vundefined_f32m8() +#define vundefined_f32mf2() \ +__builtin_rvv_vundefined_f32mf2() +#endif + +#if defined(__riscv_d) +#define vlse64_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vlse64_v_f64m1((const double *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vlse64_v_f64m2((const double *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vlse64_v_f64m4((const double *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vlse64_v_f64m8((const double *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vsse64_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsse64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsse64_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_f64m2((double *)(op0), (ptrdiff_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsse64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsse64_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_f64m4((double *)(op0), (ptrdiff_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsse64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsse64_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_f64m8((double *)(op0), (ptrdiff_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsse64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vluxei8_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f64m1((const double *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei8_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f64m2((const double *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f64m4((const double *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f64m8((const double *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f64m1((const double *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei16_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f64m2((const double *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f64m4((const double *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f64m8((const double *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f64m1((const double *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei32_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f64m2((const double *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f64m4((const double *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f64m8((const double *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f64m1((const double *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei64_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f64m2((const double *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f64m4((const double *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f64m8((const double *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei8_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f64m1((const double *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei8_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f64m2((const double *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f64m4((const double *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f64m8((const double *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f64m1((const double *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei16_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f64m2((const double *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f64m4((const double *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f64m8((const double *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f64m1((const double *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei32_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f64m2((const double *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f64m4((const double *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f64m8((const double *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f64m1((const double *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei64_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f64m2((const double *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f64m4((const double *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f64m8((const double *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmv_v_v_f64m1(op0, op1) \ +__builtin_rvv_vmv_v_v_f64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_f64m2(op0, op1) \ +__builtin_rvv_vmv_v_v_f64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_f64m4(op0, op1) \ +__builtin_rvv_vmv_v_v_f64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_f64m8(op0, op1) \ +__builtin_rvv_vmv_v_v_f64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfadd_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfadd_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfadd_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfadd_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfadd_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfadd_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfadd_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfadd_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfadd_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfadd_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfadd_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfadd_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfadd_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfadd_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfadd_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfadd_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsub_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfsub_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfsub_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfsub_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfsub_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfsub_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfsub_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfsub_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfsub_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsub_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsub_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsub_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsub_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsub_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsub_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsub_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrsub_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrsub_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrsub_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrsub_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrsub_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrsub_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrsub_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrsub_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmul_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfmul_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmul_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfmul_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmul_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfmul_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmul_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfmul_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmul_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmul_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmul_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmul_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmul_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmul_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmul_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmul_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfdiv_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfdiv_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfdiv_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfdiv_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfdiv_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfdiv_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfdiv_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfdiv_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfdiv_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrdiv_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrdiv_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrdiv_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrdiv_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrdiv_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrdiv_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrdiv_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrdiv_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmacc_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmin_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfmin_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmin_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfmin_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmin_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfmin_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmin_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfmin_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmin_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmin_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmin_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmin_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmin_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmin_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmin_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmin_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmax_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfmax_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmax_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfmax_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmax_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfmax_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmax_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfmax_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmax_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmax_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmax_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmax_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmax_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmax_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmax_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmax_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnj_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfsgnj_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnj_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnj_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnj_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnj_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnj_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnj_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnj_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfabs_v_f64m1(op0, op1) \ +__builtin_rvv_vfabs_v_f64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfabs_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfabs_v_f64m2(op0, op1) \ +__builtin_rvv_vfabs_v_f64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfabs_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfabs_v_f64m4(op0, op1) \ +__builtin_rvv_vfabs_v_f64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfabs_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfabs_v_f64m8(op0, op1) \ +__builtin_rvv_vfabs_v_f64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfabs_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vmfeq_vv_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f64m1_b64((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f64m2_b32((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f64m4_b16((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f64m8_b8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vmfeq_vf_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f64m1_b64((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfeq_vf_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfeq_vf_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f64m2_b32((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfeq_vf_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfeq_vf_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f64m4_b16((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfeq_vf_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfeq_vf_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f64m8_b8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfeq_vf_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfne_vv_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f64m1_b64((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vmfne_vv_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vmfne_vv_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f64m2_b32((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vmfne_vv_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vmfne_vv_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f64m4_b16((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vmfne_vv_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vmfne_vv_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f64m8_b8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vmfne_vv_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vmfne_vf_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f64m1_b64((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfne_vf_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfne_vf_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f64m2_b32((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfne_vf_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfne_vf_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f64m4_b16((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfne_vf_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfne_vf_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f64m8_b8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfne_vf_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vmflt_vv_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f64m1_b64((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vmflt_vv_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vmflt_vv_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f64m2_b32((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vmflt_vv_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vmflt_vv_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f64m4_b16((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vmflt_vv_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vmflt_vv_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f64m8_b8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vmflt_vv_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vmflt_vf_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f64m1_b64((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vmflt_vf_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vmflt_vf_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f64m2_b32((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vmflt_vf_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vmflt_vf_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f64m4_b16((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vmflt_vf_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vmflt_vf_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f64m8_b8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vmflt_vf_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfle_vv_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f64m1_b64((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vmfle_vv_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vmfle_vv_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f64m2_b32((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vmfle_vv_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vmfle_vv_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f64m4_b16((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vmfle_vv_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vmfle_vv_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f64m8_b8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vmfle_vv_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vmfle_vf_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f64m1_b64((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfle_vf_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfle_vf_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f64m2_b32((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfle_vf_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfle_vf_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f64m4_b16((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfle_vf_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfle_vf_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f64m8_b8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfle_vf_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfgt_vv_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f64m1_b64((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f64m2_b32((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f64m4_b16((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f64m8_b8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vmfgt_vf_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f64m1_b64((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfgt_vf_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfgt_vf_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f64m2_b32((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfgt_vf_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfgt_vf_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f64m4_b16((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfgt_vf_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfgt_vf_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f64m8_b8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfgt_vf_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfge_vv_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f64m1_b64((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vmfge_vv_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vmfge_vv_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f64m2_b32((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vmfge_vv_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vmfge_vv_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f64m4_b16((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vmfge_vv_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vmfge_vv_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f64m8_b8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vmfge_vv_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vmfge_vf_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f64m1_b64((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfge_vf_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfge_vf_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f64m2_b32((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfge_vf_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfge_vf_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f64m4_b16((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfge_vf_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfge_vf_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f64m8_b8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfge_vf_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vmerge_vvm_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f64m1((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f64m2((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f64m4((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f64m8((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmerge_vfm_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f64m1((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (size_t)(op3)) +#define vfmerge_vfm_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f64m2((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (size_t)(op3)) +#define vfmerge_vfm_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f64m4((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (size_t)(op3)) +#define vfmerge_vfm_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f64m8((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (size_t)(op3)) +#define vfmv_v_f_f64m1(op0, op1) \ +__builtin_rvv_vfmv_v_f_f64m1((double)(op0), (size_t)(op1)) +#define vfmv_v_f_f64m2(op0, op1) \ +__builtin_rvv_vfmv_v_f_f64m2((double)(op0), (size_t)(op1)) +#define vfmv_v_f_f64m4(op0, op1) \ +__builtin_rvv_vfmv_v_f_f64m4((double)(op0), (size_t)(op1)) +#define vfmv_v_f_f64m8(op0, op1) \ +__builtin_rvv_vfmv_v_f_f64m8((double)(op0), (size_t)(op1)) +#define vfredmax_vs_f64m1_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f64m1_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f64m1_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f64m1_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f64m2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f64m2_f64m1((vfloat64m1_t)(op0), (vfloat64m2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f64m2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f64m2_f64m1_m((vbool32_t)(op0), (vfloat64m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f64m4_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f64m4_f64m1((vfloat64m1_t)(op0), (vfloat64m4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f64m4_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f64m4_f64m1_m((vbool16_t)(op0), (vfloat64m1_t)(op1), (vfloat64m4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f64m8_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f64m8_f64m1((vfloat64m1_t)(op0), (vfloat64m8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f64m8_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f64m8_f64m1_m((vbool8_t)(op0), (vfloat64m1_t)(op1), (vfloat64m8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f64m1_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f64m1_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f64m1_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f64m1_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f64m2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f64m2_f64m1((vfloat64m1_t)(op0), (vfloat64m2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f64m2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f64m2_f64m1_m((vbool32_t)(op0), (vfloat64m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f64m4_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f64m4_f64m1((vfloat64m1_t)(op0), (vfloat64m4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f64m4_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f64m4_f64m1_m((vbool16_t)(op0), (vfloat64m1_t)(op1), (vfloat64m4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f64m8_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f64m8_f64m1((vfloat64m1_t)(op0), (vfloat64m8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f64m8_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f64m8_f64m1_m((vbool8_t)(op0), (vfloat64m1_t)(op1), (vfloat64m8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f64m1_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f64m1_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f64m1_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f64m1_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f64m2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f64m2_f64m1((vfloat64m1_t)(op0), (vfloat64m2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f64m2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f64m2_f64m1_m((vbool32_t)(op0), (vfloat64m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f64m4_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f64m4_f64m1((vfloat64m1_t)(op0), (vfloat64m4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f64m4_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f64m4_f64m1_m((vbool16_t)(op0), (vfloat64m1_t)(op1), (vfloat64m4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f64m8_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f64m8_f64m1((vfloat64m1_t)(op0), (vfloat64m8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f64m8_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f64m8_f64m1_m((vbool8_t)(op0), (vfloat64m1_t)(op1), (vfloat64m8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f64m1_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f64m1_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f64m1_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f64m1_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f64m2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f64m2_f64m1((vfloat64m1_t)(op0), (vfloat64m2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f64m2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f64m2_f64m1_m((vbool32_t)(op0), (vfloat64m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f64m4_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f64m4_f64m1((vfloat64m1_t)(op0), (vfloat64m4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f64m4_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f64m4_f64m1_m((vbool16_t)(op0), (vfloat64m1_t)(op1), (vfloat64m4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f64m8_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f64m8_f64m1((vfloat64m1_t)(op0), (vfloat64m8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f64m8_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f64m8_f64m1_m((vbool8_t)(op0), (vfloat64m1_t)(op1), (vfloat64m8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmv_f_s_f64m1_f64(op0) \ +__builtin_rvv_vfmv_f_s_f64m1_f64((vfloat64m1_t)(op0)) +#define vfmv_f_s_f64m2_f64(op0) \ +__builtin_rvv_vfmv_f_s_f64m2_f64((vfloat64m2_t)(op0)) +#define vfmv_f_s_f64m4_f64(op0) \ +__builtin_rvv_vfmv_f_s_f64m4_f64((vfloat64m4_t)(op0)) +#define vfmv_f_s_f64m8_f64(op0) \ +__builtin_rvv_vfmv_f_s_f64m8_f64((vfloat64m8_t)(op0)) +#define vfmv_s_f_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmv_s_f_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmv_s_f_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmv_s_f_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vslideup_vx_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vfslide1up_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1up_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1up_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1up_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1up_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1up_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1up_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1up_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1down_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1down_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1down_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1down_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1down_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1down_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1down_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1down_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vrgather_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f64m1((vfloat64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f64m2((vfloat64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f64m4((vfloat64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f64m8((vfloat64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vrgather_vx_f64m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f64m1((vfloat64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f64m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f64m2((vfloat64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f64m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f64m4((vfloat64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f64m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f64m8((vfloat64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f64m1((vfloat64m1_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f64m2((vfloat64m2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f64m4((vfloat64m4_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f64m8((vfloat64m8_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vcompress_vm_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f64m1((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f64m2((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f64m4((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f64m8((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vget_v_f64m2_f64m1(op0, op1) \ +__builtin_rvv_vget_v_f64m2_f64m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vget_v_f64m4_f64m1(op0, op1) \ +__builtin_rvv_vget_v_f64m4_f64m1((vfloat64m4_t)(op0), (size_t)(op1)) +#define vget_v_f64m8_f64m1(op0, op1) \ +__builtin_rvv_vget_v_f64m8_f64m1((vfloat64m8_t)(op0), (size_t)(op1)) +#define vget_v_f64m4_f64m2(op0, op1) \ +__builtin_rvv_vget_v_f64m4_f64m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vget_v_f64m8_f64m2(op0, op1) \ +__builtin_rvv_vget_v_f64m8_f64m2((vfloat64m8_t)(op0), (size_t)(op1)) +#define vget_v_f64m8_f64m4(op0, op1) \ +__builtin_rvv_vget_v_f64m8_f64m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vset_v_f64m1_f64m2(op0, op1, op2) \ +__builtin_rvv_vset_v_f64m1_f64m2((vfloat64m2_t)(op0), (size_t)(op1), (vfloat64m1_t)(op2)) +#define vset_v_f64m1_f64m4(op0, op1, op2) \ +__builtin_rvv_vset_v_f64m1_f64m4((vfloat64m4_t)(op0), (size_t)(op1), (vfloat64m1_t)(op2)) +#define vset_v_f64m2_f64m4(op0, op1, op2) \ +__builtin_rvv_vset_v_f64m2_f64m4((vfloat64m4_t)(op0), (size_t)(op1), (vfloat64m2_t)(op2)) +#define vset_v_f64m1_f64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_f64m1_f64m8((vfloat64m8_t)(op0), (size_t)(op1), (vfloat64m1_t)(op2)) +#define vset_v_f64m2_f64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_f64m2_f64m8((vfloat64m8_t)(op0), (size_t)(op1), (vfloat64m2_t)(op2)) +#define vset_v_f64m4_f64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_f64m4_f64m8((vfloat64m8_t)(op0), (size_t)(op1), (vfloat64m4_t)(op2)) +#define vsuxei8_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f64m4((double *)(op0), (vuint8mf2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint8mf2_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f64m8((double *)(op0), (vuint8m1_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint8m1_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f64m4((double *)(op0), (vuint16m1_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint16m1_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f64m8((double *)(op0), (vuint16m2_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint16m2_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f64m4((double *)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint32m2_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f64m8((double *)(op0), (vuint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint32m4_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f64m4((double *)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f64m8((double *)(op0), (vuint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f64m4((double *)(op0), (vuint8mf2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint8mf2_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f64m8((double *)(op0), (vuint8m1_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint8m1_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f64m4((double *)(op0), (vuint16m1_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint16m1_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f64m8((double *)(op0), (vuint16m2_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint16m2_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f64m4((double *)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint32m2_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f64m8((double *)(op0), (vuint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint32m4_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f64m4((double *)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f64m8((double *)(op0), (vuint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vle64ff_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_f64m1((const double *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_f64m2((const double *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_f64m4((const double *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_f64m8((const double *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vfneg_v_f64m1(op0, op1) \ +__builtin_rvv_vfneg_v_f64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfneg_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfneg_v_f64m2(op0, op1) \ +__builtin_rvv_vfneg_v_f64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfneg_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfneg_v_f64m4(op0, op1) \ +__builtin_rvv_vfneg_v_f64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfneg_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfneg_v_f64m8(op0, op1) \ +__builtin_rvv_vfneg_v_f64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfneg_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vle64_v_f64m1(op0, op1) \ +__builtin_rvv_vle64_v_f64m1((const double *)(op0), (size_t)(op1)) +#define vle64_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (size_t)(op3)) +#define vle64_v_f64m2(op0, op1) \ +__builtin_rvv_vle64_v_f64m2((const double *)(op0), (size_t)(op1)) +#define vle64_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (size_t)(op3)) +#define vle64_v_f64m4(op0, op1) \ +__builtin_rvv_vle64_v_f64m4((const double *)(op0), (size_t)(op1)) +#define vle64_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (size_t)(op3)) +#define vle64_v_f64m8(op0, op1) \ +__builtin_rvv_vle64_v_f64m8((const double *)(op0), (size_t)(op1)) +#define vle64_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (size_t)(op3)) +#define vse64_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vse64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vse64_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vse64_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vse64_v_f64m2((double *)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vse64_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vse64_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vse64_v_f64m4((double *)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vse64_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vse64_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vse64_v_f64m8((double *)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vse64_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfclass_v_u64m1(op0, op1) \ +__builtin_rvv_vfclass_v_u64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfclass_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfclass_v_u64m2(op0, op1) \ +__builtin_rvv_vfclass_v_u64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfclass_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfclass_v_u64m4(op0, op1) \ +__builtin_rvv_vfclass_v_u64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfclass_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfclass_v_u64m8(op0, op1) \ +__builtin_rvv_vfclass_v_u64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfclass_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f64m1(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f64m1((vint64m1_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f64m2(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f64m2((vint64m2_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f64m4(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f64m4((vint64m4_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f64m8(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f64m8((vint64m8_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f64m1(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f64m1((vuint64m1_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f64m2(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f64m2((vuint64m2_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f64m4(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f64m4((vuint64m4_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f64m8(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f64m8((vuint64m8_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i64m1(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i64m2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i64m4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i64m8(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u64m1(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u64m2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u64m4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u64m8(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i64m1(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i64m2(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i64m4(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i64m8(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u64m1(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u64m2(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u64m4(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u64m8(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i32mf2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32mf2((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i32m1(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i32m2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i32m4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u32mf2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32mf2((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u32m1(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u32m2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u32m4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i32mf2(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i32mf2((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i32m1(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i32m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i32m2(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i32m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i32m4(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i32m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u32mf2(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u32mf2((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u32m1(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u32m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u32m2(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u32m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u32m4(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u32m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfrec7_v_f64m1(op0, op1) \ +__builtin_rvv_vfrec7_v_f64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfrec7_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfrec7_v_f64m2(op0, op1) \ +__builtin_rvv_vfrec7_v_f64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfrec7_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfrec7_v_f64m4(op0, op1) \ +__builtin_rvv_vfrec7_v_f64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfrec7_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfrec7_v_f64m8(op0, op1) \ +__builtin_rvv_vfrec7_v_f64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfrec7_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f64m1(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f64m2(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f64m4(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f64m8(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f64m1(op0, op1) \ +__builtin_rvv_vfsqrt_v_f64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f64m2(op0, op1) \ +__builtin_rvv_vfsqrt_v_f64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f64m4(op0, op1) \ +__builtin_rvv_vfsqrt_v_f64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f64m8(op0, op1) \ +__builtin_rvv_vfsqrt_v_f64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f64m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f64m1((vint32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f64m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f64m2((vint32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f64m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f64m4((vint32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f64m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f64m8((vint32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f64m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m1((vuint32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f64m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m2((vuint32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f64m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m4((vuint32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f64m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m8((vuint32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vlmul_ext_v_f64m1_f64m2(op0) \ +__builtin_rvv_vlmul_ext_v_f64m1_f64m2((vfloat64m1_t)(op0)) +#define vlmul_ext_v_f64m1_f64m4(op0) \ +__builtin_rvv_vlmul_ext_v_f64m1_f64m4((vfloat64m1_t)(op0)) +#define vlmul_ext_v_f64m2_f64m4(op0) \ +__builtin_rvv_vlmul_ext_v_f64m2_f64m4((vfloat64m2_t)(op0)) +#define vlmul_ext_v_f64m1_f64m8(op0) \ +__builtin_rvv_vlmul_ext_v_f64m1_f64m8((vfloat64m1_t)(op0)) +#define vlmul_ext_v_f64m2_f64m8(op0) \ +__builtin_rvv_vlmul_ext_v_f64m2_f64m8((vfloat64m2_t)(op0)) +#define vlmul_ext_v_f64m4_f64m8(op0) \ +__builtin_rvv_vlmul_ext_v_f64m4_f64m8((vfloat64m4_t)(op0)) +#define vlmul_trunc_v_f64m2_f64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f64m2_f64m1((vfloat64m2_t)(op0)) +#define vlmul_trunc_v_f64m4_f64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f64m4_f64m1((vfloat64m4_t)(op0)) +#define vlmul_trunc_v_f64m8_f64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f64m8_f64m1((vfloat64m8_t)(op0)) +#define vlmul_trunc_v_f64m4_f64m2(op0) \ +__builtin_rvv_vlmul_trunc_v_f64m4_f64m2((vfloat64m4_t)(op0)) +#define vlmul_trunc_v_f64m8_f64m2(op0) \ +__builtin_rvv_vlmul_trunc_v_f64m8_f64m2((vfloat64m8_t)(op0)) +#define vlmul_trunc_v_f64m8_f64m4(op0) \ +__builtin_rvv_vlmul_trunc_v_f64m8_f64m4((vfloat64m8_t)(op0)) +#define vreinterpret_v_i64m1_f64m1(op0) \ +__builtin_rvv_vreinterpret_v_i64m1_f64m1((vint64m1_t)(op0)) +#define vreinterpret_v_i64m2_f64m2(op0) \ +__builtin_rvv_vreinterpret_v_i64m2_f64m2((vint64m2_t)(op0)) +#define vreinterpret_v_i64m4_f64m4(op0) \ +__builtin_rvv_vreinterpret_v_i64m4_f64m4((vint64m4_t)(op0)) +#define vreinterpret_v_i64m8_f64m8(op0) \ +__builtin_rvv_vreinterpret_v_i64m8_f64m8((vint64m8_t)(op0)) +#define vreinterpret_v_u64m1_f64m1(op0) \ +__builtin_rvv_vreinterpret_v_u64m1_f64m1((vuint64m1_t)(op0)) +#define vreinterpret_v_u64m2_f64m2(op0) \ +__builtin_rvv_vreinterpret_v_u64m2_f64m2((vuint64m2_t)(op0)) +#define vreinterpret_v_u64m4_f64m4(op0) \ +__builtin_rvv_vreinterpret_v_u64m4_f64m4((vuint64m4_t)(op0)) +#define vreinterpret_v_u64m8_f64m8(op0) \ +__builtin_rvv_vreinterpret_v_u64m8_f64m8((vuint64m8_t)(op0)) +#define vreinterpret_v_f64m1_i64m1(op0) \ +__builtin_rvv_vreinterpret_v_f64m1_i64m1((vfloat64m1_t)(op0)) +#define vreinterpret_v_f64m2_i64m2(op0) \ +__builtin_rvv_vreinterpret_v_f64m2_i64m2((vfloat64m2_t)(op0)) +#define vreinterpret_v_f64m4_i64m4(op0) \ +__builtin_rvv_vreinterpret_v_f64m4_i64m4((vfloat64m4_t)(op0)) +#define vreinterpret_v_f64m8_i64m8(op0) \ +__builtin_rvv_vreinterpret_v_f64m8_i64m8((vfloat64m8_t)(op0)) +#define vreinterpret_v_f64m1_u64m1(op0) \ +__builtin_rvv_vreinterpret_v_f64m1_u64m1((vfloat64m1_t)(op0)) +#define vreinterpret_v_f64m2_u64m2(op0) \ +__builtin_rvv_vreinterpret_v_f64m2_u64m2((vfloat64m2_t)(op0)) +#define vreinterpret_v_f64m4_u64m4(op0) \ +__builtin_rvv_vreinterpret_v_f64m4_u64m4((vfloat64m4_t)(op0)) +#define vreinterpret_v_f64m8_u64m8(op0) \ +__builtin_rvv_vreinterpret_v_f64m8_u64m8((vfloat64m8_t)(op0)) +#define vundefined_f64m1() \ +__builtin_rvv_vundefined_f64m1() +#define vundefined_f64m2() \ +__builtin_rvv_vundefined_f64m2() +#define vundefined_f64m4() \ +__builtin_rvv_vundefined_f64m4() +#define vundefined_f64m8() \ +__builtin_rvv_vundefined_f64m8() +#endif + +#if defined(__riscv_f) && defined(__riscv_d) +#define vfwadd_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f64m1((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f64m2((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f64m4((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f64m8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwadd_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f64m1((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f64m2((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f64m4((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f64m8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f64m1((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f64m2((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f64m4((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f64m8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwsub_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f64m1((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f64m2((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f64m4((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f64m8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_wv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f64m2((vfloat64m2_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f64m4((vfloat64m4_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f64m8((vfloat64m8_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwadd_wf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f64m1((vfloat64m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_wf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_wf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f64m2((vfloat64m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_wf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_wf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f64m4((vfloat64m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_wf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_wf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f64m8((vfloat64m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_wf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_wv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f64m2((vfloat64m2_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f64m4((vfloat64m4_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f64m8((vfloat64m8_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwsub_wf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f64m1((vfloat64m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_wf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_wf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f64m2((vfloat64m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_wf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_wf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f64m4((vfloat64m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_wf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_wf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f64m8((vfloat64m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_wf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwmul_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f64m1((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f64m2((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f64m4((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f64m8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwmul_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f64m1((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwmul_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwmul_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f64m2((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwmul_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwmul_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f64m4((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwmul_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwmul_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f64m8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwmul_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwmacc_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f64m2((vfloat64m2_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f64m4((vfloat64m4_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f64m8((vfloat64m8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f64m1((vfloat64m1_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f64m2((vfloat64m2_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f64m4((vfloat64m4_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f64m8((vfloat64m8_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f64m2((vfloat64m2_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f64m4((vfloat64m4_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f64m8((vfloat64m8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f64m1((vfloat64m1_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f64m2((vfloat64m2_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f64m4((vfloat64m4_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f64m8((vfloat64m8_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f64m2((vfloat64m2_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f64m4((vfloat64m4_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f64m8((vfloat64m8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f64m1((vfloat64m1_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f64m2((vfloat64m2_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f64m4((vfloat64m4_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f64m8((vfloat64m8_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f64m2((vfloat64m2_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f64m4((vfloat64m4_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f64m8((vfloat64m8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f64m1((vfloat64m1_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f64m2((vfloat64m2_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f64m4((vfloat64m4_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f64m8((vfloat64m8_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f32m1_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f32m1_f64m1((vfloat64m1_t)(op0), (vfloat32m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f32m1_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f32m1_f64m1_m((vbool32_t)(op0), (vfloat64m1_t)(op1), (vfloat32m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f32m2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f32m2_f64m1((vfloat64m1_t)(op0), (vfloat32m2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f32m2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f32m2_f64m1_m((vbool16_t)(op0), (vfloat64m1_t)(op1), (vfloat32m2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f32m4_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f32m4_f64m1((vfloat64m1_t)(op0), (vfloat32m4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f32m4_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f32m4_f64m1_m((vbool8_t)(op0), (vfloat64m1_t)(op1), (vfloat32m4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f32m8_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f32m8_f64m1((vfloat64m1_t)(op0), (vfloat32m8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f32m8_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f32m8_f64m1_m((vbool4_t)(op0), (vfloat64m1_t)(op1), (vfloat32m8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f32mf2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f32mf2_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f32mf2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f32mf2_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f32m1_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f32m1_f64m1((vfloat64m1_t)(op0), (vfloat32m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f32m1_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f32m1_f64m1_m((vbool32_t)(op0), (vfloat64m1_t)(op1), (vfloat32m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f32m2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f32m2_f64m1((vfloat64m1_t)(op0), (vfloat32m2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f32m2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f32m2_f64m1_m((vbool16_t)(op0), (vfloat64m1_t)(op1), (vfloat32m2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f32m4_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f32m4_f64m1((vfloat64m1_t)(op0), (vfloat32m4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f32m4_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f32m4_f64m1_m((vbool8_t)(op0), (vfloat64m1_t)(op1), (vfloat32m4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f32m8_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f32m8_f64m1((vfloat64m1_t)(op0), (vfloat32m8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f32m8_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f32m8_f64m1_m((vbool4_t)(op0), (vfloat64m1_t)(op1), (vfloat32m8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f32mf2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f32mf2_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f32mf2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f32mf2_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfncvt_f_f_w_f32mf2(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f32mf2((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f32m1(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f32m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f32m2(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f32m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f32m4(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f32m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f32mf2(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32mf2((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f32m1(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f32m2(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f32m4(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f64m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f64m1((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f64m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f64m2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f64m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f64m4((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f64m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f64m8((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#endif + +#if defined(__riscv_zfh) +#define vsse16_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsse16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsse16_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_f16m2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsse16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsse16_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_f16m4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsse16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsse16_v_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_f16m8((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vsse16_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_f16m8_m((vbool2_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vsse16_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsse16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsse16_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsse16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f16m1((const _Float16 *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f16m2((const _Float16 *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f16m4((const _Float16 *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f16m8((const _Float16 *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vluxei8_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxei8_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f16mf2((const _Float16 *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f16mf4((const _Float16 *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei16_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f16m1((const _Float16 *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f16m2((const _Float16 *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f16m4((const _Float16 *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f16m8((const _Float16 *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vluxei16_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxei16_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f16mf2((const _Float16 *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f16mf4((const _Float16 *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei32_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f16m1((const _Float16 *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f16m2((const _Float16 *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f16m4((const _Float16 *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f16mf2((const _Float16 *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f16mf4((const _Float16 *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei64_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f16m1((const _Float16 *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f16m2((const _Float16 *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f16mf2((const _Float16 *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f16mf4((const _Float16 *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f16m1((const _Float16 *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f16m2((const _Float16 *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f16m4((const _Float16 *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f16m8((const _Float16 *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vloxei8_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxei8_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f16mf2((const _Float16 *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f16mf4((const _Float16 *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei16_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f16m1((const _Float16 *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f16m2((const _Float16 *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f16m4((const _Float16 *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f16m8((const _Float16 *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vloxei16_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxei16_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f16mf2((const _Float16 *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f16mf4((const _Float16 *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei32_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f16m1((const _Float16 *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f16m2((const _Float16 *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f16m4((const _Float16 *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f16mf2((const _Float16 *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f16mf4((const _Float16 *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei64_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f16m1((const _Float16 *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f16m2((const _Float16 *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f16mf2((const _Float16 *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f16mf4((const _Float16 *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmv_v_v_f16m1(op0, op1) \ +__builtin_rvv_vmv_v_v_f16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_f16m2(op0, op1) \ +__builtin_rvv_vmv_v_v_f16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_f16m4(op0, op1) \ +__builtin_rvv_vmv_v_v_f16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_f16m8(op0, op1) \ +__builtin_rvv_vmv_v_v_f16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_f16mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_f16mf4(op0, op1) \ +__builtin_rvv_vmv_v_v_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfadd_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfadd_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfadd_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfadd_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfadd_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfadd_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfadd_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfadd_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfadd_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfadd_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfadd_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfadd_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfadd_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfadd_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfadd_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfadd_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfadd_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfadd_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfadd_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfadd_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfadd_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfadd_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfadd_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfadd_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsub_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfsub_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfsub_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfsub_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfsub_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfsub_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfsub_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfsub_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfsub_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfsub_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfsub_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfsub_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfsub_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsub_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsub_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsub_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsub_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsub_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsub_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsub_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsub_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsub_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsub_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsub_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrsub_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrsub_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrsub_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrsub_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrsub_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrsub_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrsub_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrsub_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrsub_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrsub_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrsub_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrsub_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmul_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfmul_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmul_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfmul_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmul_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfmul_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmul_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfmul_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmul_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfmul_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmul_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfmul_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmul_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmul_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmul_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmul_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmul_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmul_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmul_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmul_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmul_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmul_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmul_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmul_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfdiv_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfdiv_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfdiv_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfdiv_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfdiv_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfdiv_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfdiv_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfdiv_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfdiv_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfdiv_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfdiv_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfdiv_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfdiv_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrdiv_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrdiv_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrdiv_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrdiv_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrdiv_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrdiv_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrdiv_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrdiv_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrdiv_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrdiv_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrdiv_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrdiv_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmacc_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmin_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfmin_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmin_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfmin_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmin_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfmin_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmin_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfmin_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmin_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfmin_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmin_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfmin_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmin_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmin_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmin_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmin_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmin_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmin_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmin_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmin_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmin_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmin_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmin_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmin_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmax_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfmax_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmax_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfmax_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmax_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfmax_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmax_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfmax_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmax_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfmax_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmax_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfmax_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmax_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmax_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmax_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmax_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmax_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmax_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmax_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmax_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmax_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmax_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmax_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmax_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnj_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfsgnj_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnj_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnj_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnj_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnj_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnj_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnj_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnj_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnj_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnj_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnj_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnj_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfabs_v_f16m1(op0, op1) \ +__builtin_rvv_vfabs_v_f16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfabs_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfabs_v_f16m2(op0, op1) \ +__builtin_rvv_vfabs_v_f16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfabs_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfabs_v_f16m4(op0, op1) \ +__builtin_rvv_vfabs_v_f16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfabs_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfabs_v_f16m8(op0, op1) \ +__builtin_rvv_vfabs_v_f16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfabs_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfabs_v_f16mf2(op0, op1) \ +__builtin_rvv_vfabs_v_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfabs_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfabs_v_f16mf4(op0, op1) \ +__builtin_rvv_vfabs_v_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfabs_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vmfeq_vv_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f16m1_b16((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f16m2_b8((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f16m4_b4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f16m8_b2((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f16mf2_b32((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f16mf4_b64((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vmfeq_vf_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f16m1_b16((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfeq_vf_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfeq_vf_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f16m2_b8((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfeq_vf_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfeq_vf_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f16m4_b4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfeq_vf_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfeq_vf_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f16m8_b2((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfeq_vf_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfeq_vf_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f16mf2_b32((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfeq_vf_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfeq_vf_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f16mf4_b64((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfeq_vf_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfne_vv_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f16m1_b16((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vmfne_vv_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vmfne_vv_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f16m2_b8((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vmfne_vv_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vmfne_vv_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f16m4_b4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vmfne_vv_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vmfne_vv_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f16m8_b2((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vmfne_vv_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vmfne_vv_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f16mf2_b32((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vmfne_vv_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vmfne_vv_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f16mf4_b64((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vmfne_vv_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vmfne_vf_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f16m1_b16((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfne_vf_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfne_vf_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f16m2_b8((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfne_vf_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfne_vf_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f16m4_b4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfne_vf_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfne_vf_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f16m8_b2((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfne_vf_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfne_vf_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f16mf2_b32((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfne_vf_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfne_vf_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f16mf4_b64((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfne_vf_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmflt_vv_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f16m1_b16((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vmflt_vv_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vmflt_vv_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f16m2_b8((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vmflt_vv_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vmflt_vv_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f16m4_b4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vmflt_vv_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vmflt_vv_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f16m8_b2((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vmflt_vv_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vmflt_vv_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f16mf2_b32((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vmflt_vv_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vmflt_vv_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f16mf4_b64((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vmflt_vv_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vmflt_vf_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f16m1_b16((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmflt_vf_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmflt_vf_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f16m2_b8((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmflt_vf_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmflt_vf_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f16m4_b4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmflt_vf_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmflt_vf_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f16m8_b2((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmflt_vf_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmflt_vf_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f16mf2_b32((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmflt_vf_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmflt_vf_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f16mf4_b64((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmflt_vf_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfle_vv_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f16m1_b16((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vmfle_vv_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vmfle_vv_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f16m2_b8((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vmfle_vv_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vmfle_vv_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f16m4_b4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vmfle_vv_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vmfle_vv_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f16m8_b2((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vmfle_vv_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vmfle_vv_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f16mf2_b32((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vmfle_vv_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vmfle_vv_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f16mf4_b64((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vmfle_vv_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vmfle_vf_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f16m1_b16((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfle_vf_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfle_vf_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f16m2_b8((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfle_vf_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfle_vf_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f16m4_b4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfle_vf_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfle_vf_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f16m8_b2((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfle_vf_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfle_vf_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f16mf2_b32((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfle_vf_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfle_vf_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f16mf4_b64((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfle_vf_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfgt_vv_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f16m1_b16((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f16m2_b8((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f16m4_b4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f16m8_b2((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f16mf2_b32((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f16mf4_b64((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vmfgt_vf_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f16m1_b16((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfgt_vf_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfgt_vf_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f16m2_b8((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfgt_vf_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfgt_vf_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f16m4_b4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfgt_vf_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfgt_vf_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f16m8_b2((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfgt_vf_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfgt_vf_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f16mf2_b32((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfgt_vf_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfgt_vf_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f16mf4_b64((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfgt_vf_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfge_vv_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f16m1_b16((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vmfge_vv_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vmfge_vv_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f16m2_b8((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vmfge_vv_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vmfge_vv_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f16m4_b4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vmfge_vv_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vmfge_vv_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f16m8_b2((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vmfge_vv_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vmfge_vv_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f16mf2_b32((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vmfge_vv_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vmfge_vv_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f16mf4_b64((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vmfge_vv_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vmfge_vf_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f16m1_b16((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfge_vf_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfge_vf_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f16m2_b8((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfge_vf_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfge_vf_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f16m4_b4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfge_vf_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfge_vf_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f16m8_b2((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfge_vf_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfge_vf_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f16mf2_b32((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfge_vf_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfge_vf_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f16mf4_b64((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfge_vf_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmerge_vvm_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f16m1((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f16m2((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f16m4((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f16m8((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f16mf2((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f16mf4((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmerge_vfm_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f16m1((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (size_t)(op3)) +#define vfmerge_vfm_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f16m2((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (size_t)(op3)) +#define vfmerge_vfm_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f16m4((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (size_t)(op3)) +#define vfmerge_vfm_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f16m8((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (size_t)(op3)) +#define vfmerge_vfm_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f16mf2((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (size_t)(op3)) +#define vfmerge_vfm_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f16mf4((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (size_t)(op3)) +#define vfmv_v_f_f16m1(op0, op1) \ +__builtin_rvv_vfmv_v_f_f16m1((_Float16)(op0), (size_t)(op1)) +#define vfmv_v_f_f16m2(op0, op1) \ +__builtin_rvv_vfmv_v_f_f16m2((_Float16)(op0), (size_t)(op1)) +#define vfmv_v_f_f16m4(op0, op1) \ +__builtin_rvv_vfmv_v_f_f16m4((_Float16)(op0), (size_t)(op1)) +#define vfmv_v_f_f16m8(op0, op1) \ +__builtin_rvv_vfmv_v_f_f16m8((_Float16)(op0), (size_t)(op1)) +#define vfmv_v_f_f16mf2(op0, op1) \ +__builtin_rvv_vfmv_v_f_f16mf2((_Float16)(op0), (size_t)(op1)) +#define vfmv_v_f_f16mf4(op0, op1) \ +__builtin_rvv_vfmv_v_f_f16mf4((_Float16)(op0), (size_t)(op1)) +#define vfredmax_vs_f16m1_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f16m1_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f16m1_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f16m1_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f16m2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f16m2_f16m1((vfloat16m1_t)(op0), (vfloat16m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f16m2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f16m2_f16m1_m((vbool8_t)(op0), (vfloat16m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f16m4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f16m4_f16m1((vfloat16m1_t)(op0), (vfloat16m4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f16m4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f16m4_f16m1_m((vbool4_t)(op0), (vfloat16m1_t)(op1), (vfloat16m4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f16m8_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f16m8_f16m1((vfloat16m1_t)(op0), (vfloat16m8_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f16m8_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f16m8_f16m1_m((vbool2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m8_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f16mf2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f16mf2_f16m1((vfloat16m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f16mf2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f16mf2_f16m1_m((vbool32_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f16mf4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f16mf4_f16m1((vfloat16m1_t)(op0), (vfloat16mf4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f16mf4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f16mf4_f16m1_m((vbool64_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f16m1_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f16m1_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f16m1_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f16m1_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f16m2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f16m2_f16m1((vfloat16m1_t)(op0), (vfloat16m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f16m2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f16m2_f16m1_m((vbool8_t)(op0), (vfloat16m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f16m4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f16m4_f16m1((vfloat16m1_t)(op0), (vfloat16m4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f16m4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f16m4_f16m1_m((vbool4_t)(op0), (vfloat16m1_t)(op1), (vfloat16m4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f16m8_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f16m8_f16m1((vfloat16m1_t)(op0), (vfloat16m8_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f16m8_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f16m8_f16m1_m((vbool2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m8_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f16mf2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f16mf2_f16m1((vfloat16m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f16mf2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f16mf2_f16m1_m((vbool32_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f16mf4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f16mf4_f16m1((vfloat16m1_t)(op0), (vfloat16mf4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f16mf4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f16mf4_f16m1_m((vbool64_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f16m1_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f16m1_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f16m1_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f16m1_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f16m2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f16m2_f16m1((vfloat16m1_t)(op0), (vfloat16m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f16m2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f16m2_f16m1_m((vbool8_t)(op0), (vfloat16m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f16m4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f16m4_f16m1((vfloat16m1_t)(op0), (vfloat16m4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f16m4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f16m4_f16m1_m((vbool4_t)(op0), (vfloat16m1_t)(op1), (vfloat16m4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f16m8_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f16m8_f16m1((vfloat16m1_t)(op0), (vfloat16m8_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f16m8_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f16m8_f16m1_m((vbool2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m8_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f16mf2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f16mf2_f16m1((vfloat16m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f16mf2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f16mf2_f16m1_m((vbool32_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f16mf4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f16mf4_f16m1((vfloat16m1_t)(op0), (vfloat16mf4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f16mf4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f16mf4_f16m1_m((vbool64_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f16m1_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f16m1_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f16m1_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f16m1_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f16m2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f16m2_f16m1((vfloat16m1_t)(op0), (vfloat16m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f16m2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f16m2_f16m1_m((vbool8_t)(op0), (vfloat16m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f16m4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f16m4_f16m1((vfloat16m1_t)(op0), (vfloat16m4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f16m4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f16m4_f16m1_m((vbool4_t)(op0), (vfloat16m1_t)(op1), (vfloat16m4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f16m8_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f16m8_f16m1((vfloat16m1_t)(op0), (vfloat16m8_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f16m8_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f16m8_f16m1_m((vbool2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m8_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f16mf2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f16mf2_f16m1((vfloat16m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f16mf2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f16mf2_f16m1_m((vbool32_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f16mf4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f16mf4_f16m1((vfloat16m1_t)(op0), (vfloat16mf4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f16mf4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f16mf4_f16m1_m((vbool64_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f16m4((_Float16 *)(op0), (vuint8m2_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint8m2_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f16m8((_Float16 *)(op0), (vuint8m4_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f16m8_m((vbool2_t)(op0), (_Float16 *)(op1), (vuint8m4_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f16m4((_Float16 *)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f16m8((_Float16 *)(op0), (vuint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f16m8_m((vbool2_t)(op0), (_Float16 *)(op1), (vuint16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmv_f_s_f16m1_f16(op0) \ +__builtin_rvv_vfmv_f_s_f16m1_f16((vfloat16m1_t)(op0)) +#define vfmv_f_s_f16m2_f16(op0) \ +__builtin_rvv_vfmv_f_s_f16m2_f16((vfloat16m2_t)(op0)) +#define vfmv_f_s_f16m4_f16(op0) \ +__builtin_rvv_vfmv_f_s_f16m4_f16((vfloat16m4_t)(op0)) +#define vfmv_f_s_f16m8_f16(op0) \ +__builtin_rvv_vfmv_f_s_f16m8_f16((vfloat16m8_t)(op0)) +#define vfmv_f_s_f16mf2_f16(op0) \ +__builtin_rvv_vfmv_f_s_f16mf2_f16((vfloat16mf2_t)(op0)) +#define vfmv_f_s_f16mf4_f16(op0) \ +__builtin_rvv_vfmv_f_s_f16mf4_f16((vfloat16mf4_t)(op0)) +#define vfmv_s_f_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmv_s_f_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmv_s_f_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmv_s_f_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmv_s_f_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmv_s_f_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vslideup_vx_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vfslide1up_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1up_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1up_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1up_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1up_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1up_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1up_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1up_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1up_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1up_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1up_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1up_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vsuxei32_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f16m4((_Float16 *)(op0), (vuint32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint32m8_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfslide1down_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1down_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1down_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1down_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1down_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1down_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1down_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1down_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1down_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1down_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1down_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1down_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vrgather_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f16m1((vfloat16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f16m2((vfloat16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f16m4((vfloat16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f16m8((vfloat16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f16mf2((vfloat16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f16mf4((vfloat16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgather_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgather_vx_f16m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f16m1((vfloat16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f16m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f16m2((vfloat16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f16m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f16m4((vfloat16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f16m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f16m8((vfloat16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f16mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f16mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f16m1((vfloat16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f16m2((vfloat16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f16m4((vfloat16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f16m8((vfloat16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f16mf2((vfloat16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f16mf4((vfloat16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vcompress_vm_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f16m1((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f16m2((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f16m4((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f16m8((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f16mf2((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f16mf4((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f16m4((_Float16 *)(op0), (vuint8m2_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint8m2_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f16m8((_Float16 *)(op0), (vuint8m4_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f16m8_m((vbool2_t)(op0), (_Float16 *)(op1), (vuint8m4_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f16m4((_Float16 *)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f16m8((_Float16 *)(op0), (vuint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f16m8_m((vbool2_t)(op0), (_Float16 *)(op1), (vuint16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f16m4((_Float16 *)(op0), (vuint32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint32m8_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vle16ff_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_f16m1((const _Float16 *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_f16m2((const _Float16 *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_f16m4((const _Float16 *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_f16m8((const _Float16 *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_f16mf2((const _Float16 *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_f16mf4((const _Float16 *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vfneg_v_f16m1(op0, op1) \ +__builtin_rvv_vfneg_v_f16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfneg_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfneg_v_f16m2(op0, op1) \ +__builtin_rvv_vfneg_v_f16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfneg_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfneg_v_f16m4(op0, op1) \ +__builtin_rvv_vfneg_v_f16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfneg_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfneg_v_f16m8(op0, op1) \ +__builtin_rvv_vfneg_v_f16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfneg_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfneg_v_f16mf2(op0, op1) \ +__builtin_rvv_vfneg_v_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfneg_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfneg_v_f16mf4(op0, op1) \ +__builtin_rvv_vfneg_v_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfneg_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vle16_v_f16m1(op0, op1) \ +__builtin_rvv_vle16_v_f16m1((const _Float16 *)(op0), (size_t)(op1)) +#define vle16_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vle16_v_f16m2(op0, op1) \ +__builtin_rvv_vle16_v_f16m2((const _Float16 *)(op0), (size_t)(op1)) +#define vle16_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vle16_v_f16m4(op0, op1) \ +__builtin_rvv_vle16_v_f16m4((const _Float16 *)(op0), (size_t)(op1)) +#define vle16_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vle16_v_f16m8(op0, op1) \ +__builtin_rvv_vle16_v_f16m8((const _Float16 *)(op0), (size_t)(op1)) +#define vle16_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vle16_v_f16mf2(op0, op1) \ +__builtin_rvv_vle16_v_f16mf2((const _Float16 *)(op0), (size_t)(op1)) +#define vle16_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vle16_v_f16mf4(op0, op1) \ +__builtin_rvv_vle16_v_f16mf4((const _Float16 *)(op0), (size_t)(op1)) +#define vle16_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vse16_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vse16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vse16_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vse16_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vse16_v_f16m2((_Float16 *)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vse16_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vse16_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vse16_v_f16m4((_Float16 *)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vse16_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vse16_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vse16_v_f16m8((_Float16 *)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vse16_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_f16m8_m((vbool2_t)(op0), (_Float16 *)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vse16_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vse16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vse16_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vse16_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vse16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vse16_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vlse16_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vlse16_v_f16m1((const _Float16 *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vlse16_v_f16m2((const _Float16 *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vlse16_v_f16m4((const _Float16 *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vlse16_v_f16m8((const _Float16 *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vlse16_v_f16mf2((const _Float16 *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vlse16_v_f16mf4((const _Float16 *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vfclass_v_u16m1(op0, op1) \ +__builtin_rvv_vfclass_v_u16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfclass_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfclass_v_u16m2(op0, op1) \ +__builtin_rvv_vfclass_v_u16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfclass_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfclass_v_u16m4(op0, op1) \ +__builtin_rvv_vfclass_v_u16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfclass_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfclass_v_u16m8(op0, op1) \ +__builtin_rvv_vfclass_v_u16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfclass_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfclass_v_u16mf2(op0, op1) \ +__builtin_rvv_vfclass_v_u16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfclass_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfclass_v_u16mf4(op0, op1) \ +__builtin_rvv_vfclass_v_u16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfclass_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f16m1(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f16m1((vint16m1_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f16m2(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f16m2((vint16m2_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f16m4(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f16m4((vint16m4_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f16m8(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f16m8((vint16m8_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f16mf2(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f16mf2((vint16mf2_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f16mf4(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f16mf4((vint16mf4_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f16m1(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f16m1((vuint16m1_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f16m2(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f16m2((vuint16m2_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f16m4(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f16m4((vuint16m4_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f16m8(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f16m8((vuint16m8_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f16mf2(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f16mf2((vuint16mf2_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f16mf4(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f16mf4((vuint16mf4_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i16m1(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i16m2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i16m4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i16m8(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i16mf2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i16mf4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u16m1(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u16m2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u16m4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u16m8(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u16mf2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u16mf4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i16m1(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i16m2(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i16m4(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i16m8(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i16mf2(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i16mf4(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u16m1(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u16m2(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u16m4(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u16m8(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u16mf2(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u16mf4(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f16mf4(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f16mf4((vint32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f16mf2(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f16mf2((vint32m1_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f16m1(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f16m1((vint32m2_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f16m2(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f16m2((vint32m4_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f16m4(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f16m4((vint32m8_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f16mf4(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f16mf4((vuint32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f16mf2(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f16mf2((vuint32m1_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f16m1(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f16m1((vuint32m2_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f16m2(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f16m2((vuint32m4_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f16m4(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f16m4((vuint32m8_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i8mf8(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8mf8((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i8mf4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8mf4((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i8mf2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8mf2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i8m1(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8m1((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i8m2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8m2((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i8m4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8m4((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u8mf8(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf8((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u8mf4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf4((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u8mf2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u8m1(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8m1((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u8m2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8m2((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u8m4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8m4((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i8mf8(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i8mf8((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i8mf4(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i8mf4((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i8mf2(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i8mf2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i8m1(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i8m1((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i8m2(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i8m2((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i8m4(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i8m4((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u8mf8(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u8mf8((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u8mf4(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u8mf4((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u8mf2(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u8mf2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u8m1(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u8m1((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u8m2(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u8m2((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u8m4(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u8m4((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfrec7_v_f16m1(op0, op1) \ +__builtin_rvv_vfrec7_v_f16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfrec7_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfrec7_v_f16m2(op0, op1) \ +__builtin_rvv_vfrec7_v_f16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfrec7_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfrec7_v_f16m4(op0, op1) \ +__builtin_rvv_vfrec7_v_f16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfrec7_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfrec7_v_f16m8(op0, op1) \ +__builtin_rvv_vfrec7_v_f16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfrec7_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfrec7_v_f16mf2(op0, op1) \ +__builtin_rvv_vfrec7_v_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfrec7_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfrec7_v_f16mf4(op0, op1) \ +__builtin_rvv_vfrec7_v_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfrec7_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f16m1(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f16m2(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f16m4(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f16m8(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f16mf2(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f16mf4(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f16m1(op0, op1) \ +__builtin_rvv_vfsqrt_v_f16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f16m2(op0, op1) \ +__builtin_rvv_vfsqrt_v_f16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f16m4(op0, op1) \ +__builtin_rvv_vfsqrt_v_f16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f16m8(op0, op1) \ +__builtin_rvv_vfsqrt_v_f16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f16mf2(op0, op1) \ +__builtin_rvv_vfsqrt_v_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f16mf4(op0, op1) \ +__builtin_rvv_vfsqrt_v_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f16mf4(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f16mf4((vint8mf8_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f16mf2(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f16mf2((vint8mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f16m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f16m1((vint8mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f16m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f16m2((vint8m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f16m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f16m4((vint8m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f16m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f16m8((vint8m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f16mf4(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f16mf4((vuint8mf8_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f16mf2(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f16mf2((vuint8mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f16m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m1((vuint8mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f16m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m2((vuint8m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f16m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m4((vuint8m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f16m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m8((vuint8m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32mf2((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i32m1(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m1((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i32m2(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i32m4(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m4((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i32m8(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m8((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32mf2((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u32m1(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m1((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u32m2(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u32m4(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m4((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u32m8(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m8((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i32mf2((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i32m1(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i32m1((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i32m2(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i32m2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i32m4(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i32m4((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i32m8(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i32m8((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u32mf2((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u32m1(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m1((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u32m2(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u32m4(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m4((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u32m8(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m8((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vlmul_ext_v_f16mf2_f16m1(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf2_f16m1((vfloat16mf2_t)(op0)) +#define vlmul_ext_v_f16mf4_f16m1(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf4_f16m1((vfloat16mf4_t)(op0)) +#define vlmul_ext_v_f16mf4_f16mf2(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf4_f16mf2((vfloat16mf4_t)(op0)) +#define vlmul_ext_v_f16m1_f16m2(op0) \ +__builtin_rvv_vlmul_ext_v_f16m1_f16m2((vfloat16m1_t)(op0)) +#define vlmul_ext_v_f16mf2_f16m2(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf2_f16m2((vfloat16mf2_t)(op0)) +#define vlmul_ext_v_f16mf4_f16m2(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf4_f16m2((vfloat16mf4_t)(op0)) +#define vlmul_ext_v_f16m1_f16m4(op0) \ +__builtin_rvv_vlmul_ext_v_f16m1_f16m4((vfloat16m1_t)(op0)) +#define vlmul_ext_v_f16m2_f16m4(op0) \ +__builtin_rvv_vlmul_ext_v_f16m2_f16m4((vfloat16m2_t)(op0)) +#define vlmul_ext_v_f16mf2_f16m4(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf2_f16m4((vfloat16mf2_t)(op0)) +#define vlmul_ext_v_f16mf4_f16m4(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf4_f16m4((vfloat16mf4_t)(op0)) +#define vlmul_ext_v_f16m1_f16m8(op0) \ +__builtin_rvv_vlmul_ext_v_f16m1_f16m8((vfloat16m1_t)(op0)) +#define vlmul_ext_v_f16m2_f16m8(op0) \ +__builtin_rvv_vlmul_ext_v_f16m2_f16m8((vfloat16m2_t)(op0)) +#define vlmul_ext_v_f16m4_f16m8(op0) \ +__builtin_rvv_vlmul_ext_v_f16m4_f16m8((vfloat16m4_t)(op0)) +#define vlmul_ext_v_f16mf2_f16m8(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf2_f16m8((vfloat16mf2_t)(op0)) +#define vlmul_ext_v_f16mf4_f16m8(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf4_f16m8((vfloat16mf4_t)(op0)) +#define vlmul_trunc_v_f16m1_f16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m1_f16mf2((vfloat16m1_t)(op0)) +#define vlmul_trunc_v_f16m2_f16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m2_f16mf2((vfloat16m2_t)(op0)) +#define vlmul_trunc_v_f16m4_f16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m4_f16mf2((vfloat16m4_t)(op0)) +#define vlmul_trunc_v_f16m8_f16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m8_f16mf2((vfloat16m8_t)(op0)) +#define vlmul_trunc_v_f16m1_f16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m1_f16mf4((vfloat16m1_t)(op0)) +#define vlmul_trunc_v_f16m2_f16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m2_f16mf4((vfloat16m2_t)(op0)) +#define vlmul_trunc_v_f16m4_f16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m4_f16mf4((vfloat16m4_t)(op0)) +#define vlmul_trunc_v_f16m8_f16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m8_f16mf4((vfloat16m8_t)(op0)) +#define vlmul_trunc_v_f16mf2_f16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_f16mf2_f16mf4((vfloat16mf2_t)(op0)) +#define vlmul_trunc_v_f16m2_f16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m2_f16m1((vfloat16m2_t)(op0)) +#define vlmul_trunc_v_f16m4_f16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m4_f16m1((vfloat16m4_t)(op0)) +#define vlmul_trunc_v_f16m8_f16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m8_f16m1((vfloat16m8_t)(op0)) +#define vlmul_trunc_v_f16m4_f16m2(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m4_f16m2((vfloat16m4_t)(op0)) +#define vlmul_trunc_v_f16m8_f16m2(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m8_f16m2((vfloat16m8_t)(op0)) +#define vlmul_trunc_v_f16m8_f16m4(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m8_f16m4((vfloat16m8_t)(op0)) +#define vreinterpret_v_i16m1_f16m1(op0) \ +__builtin_rvv_vreinterpret_v_i16m1_f16m1((vint16m1_t)(op0)) +#define vreinterpret_v_i16m2_f16m2(op0) \ +__builtin_rvv_vreinterpret_v_i16m2_f16m2((vint16m2_t)(op0)) +#define vreinterpret_v_i16m4_f16m4(op0) \ +__builtin_rvv_vreinterpret_v_i16m4_f16m4((vint16m4_t)(op0)) +#define vreinterpret_v_i16m8_f16m8(op0) \ +__builtin_rvv_vreinterpret_v_i16m8_f16m8((vint16m8_t)(op0)) +#define vreinterpret_v_i16mf2_f16mf2(op0) \ +__builtin_rvv_vreinterpret_v_i16mf2_f16mf2((vint16mf2_t)(op0)) +#define vreinterpret_v_i16mf4_f16mf4(op0) \ +__builtin_rvv_vreinterpret_v_i16mf4_f16mf4((vint16mf4_t)(op0)) +#define vreinterpret_v_u16m1_f16m1(op0) \ +__builtin_rvv_vreinterpret_v_u16m1_f16m1((vuint16m1_t)(op0)) +#define vreinterpret_v_u16m2_f16m2(op0) \ +__builtin_rvv_vreinterpret_v_u16m2_f16m2((vuint16m2_t)(op0)) +#define vreinterpret_v_u16m4_f16m4(op0) \ +__builtin_rvv_vreinterpret_v_u16m4_f16m4((vuint16m4_t)(op0)) +#define vreinterpret_v_u16m8_f16m8(op0) \ +__builtin_rvv_vreinterpret_v_u16m8_f16m8((vuint16m8_t)(op0)) +#define vreinterpret_v_u16mf2_f16mf2(op0) \ +__builtin_rvv_vreinterpret_v_u16mf2_f16mf2((vuint16mf2_t)(op0)) +#define vreinterpret_v_u16mf4_f16mf4(op0) \ +__builtin_rvv_vreinterpret_v_u16mf4_f16mf4((vuint16mf4_t)(op0)) +#define vreinterpret_v_f16m1_i16m1(op0) \ +__builtin_rvv_vreinterpret_v_f16m1_i16m1((vfloat16m1_t)(op0)) +#define vreinterpret_v_f16m2_i16m2(op0) \ +__builtin_rvv_vreinterpret_v_f16m2_i16m2((vfloat16m2_t)(op0)) +#define vreinterpret_v_f16m4_i16m4(op0) \ +__builtin_rvv_vreinterpret_v_f16m4_i16m4((vfloat16m4_t)(op0)) +#define vreinterpret_v_f16m8_i16m8(op0) \ +__builtin_rvv_vreinterpret_v_f16m8_i16m8((vfloat16m8_t)(op0)) +#define vreinterpret_v_f16mf2_i16mf2(op0) \ +__builtin_rvv_vreinterpret_v_f16mf2_i16mf2((vfloat16mf2_t)(op0)) +#define vreinterpret_v_f16mf4_i16mf4(op0) \ +__builtin_rvv_vreinterpret_v_f16mf4_i16mf4((vfloat16mf4_t)(op0)) +#define vreinterpret_v_f16m1_u16m1(op0) \ +__builtin_rvv_vreinterpret_v_f16m1_u16m1((vfloat16m1_t)(op0)) +#define vreinterpret_v_f16m2_u16m2(op0) \ +__builtin_rvv_vreinterpret_v_f16m2_u16m2((vfloat16m2_t)(op0)) +#define vreinterpret_v_f16m4_u16m4(op0) \ +__builtin_rvv_vreinterpret_v_f16m4_u16m4((vfloat16m4_t)(op0)) +#define vreinterpret_v_f16m8_u16m8(op0) \ +__builtin_rvv_vreinterpret_v_f16m8_u16m8((vfloat16m8_t)(op0)) +#define vreinterpret_v_f16mf2_u16mf2(op0) \ +__builtin_rvv_vreinterpret_v_f16mf2_u16mf2((vfloat16mf2_t)(op0)) +#define vreinterpret_v_f16mf4_u16mf4(op0) \ +__builtin_rvv_vreinterpret_v_f16mf4_u16mf4((vfloat16mf4_t)(op0)) +#define vundefined_f16m1() \ +__builtin_rvv_vundefined_f16m1() +#define vundefined_f16m2() \ +__builtin_rvv_vundefined_f16m2() +#define vundefined_f16m4() \ +__builtin_rvv_vundefined_f16m4() +#define vundefined_f16m8() \ +__builtin_rvv_vundefined_f16m8() +#define vundefined_f16mf2() \ +__builtin_rvv_vundefined_f16mf2() +#define vundefined_f16mf4() \ +__builtin_rvv_vundefined_f16mf4() +#endif + +#if defined(__riscv_f) && defined(__riscv_zfh) +#define vfwadd_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f32mf2((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f32m1((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f32m2((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f32m4((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f32m8((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwadd_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f32mf2((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f32m1((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f32m2((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f32m4((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f32m8((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f32mf2((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f32m1((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f32m2((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f32m4((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f32m8((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwsub_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f32mf2((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f32m1((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f32m2((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f32m4((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f32m8((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_wv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f32mf2((vfloat32mf2_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f32m2((vfloat32m2_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f32m4((vfloat32m4_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f32m8((vfloat32m8_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwadd_wf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f32mf2((vfloat32mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_wf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_wf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f32m1((vfloat32m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_wf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_wf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f32m2((vfloat32m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_wf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_wf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f32m4((vfloat32m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_wf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_wf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f32m8((vfloat32m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_wf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_wv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f32mf2((vfloat32mf2_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f32m2((vfloat32m2_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f32m4((vfloat32m4_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f32m8((vfloat32m8_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwsub_wf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f32mf2((vfloat32mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_wf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_wf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f32m1((vfloat32m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_wf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_wf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f32m2((vfloat32m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_wf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_wf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f32m4((vfloat32m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_wf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_wf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f32m8((vfloat32m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_wf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwmul_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f32mf2((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f32m1((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f32m2((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f32m4((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f32m8((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwmul_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f32mf2((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwmul_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwmul_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f32m1((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwmul_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwmul_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f32m2((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwmul_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwmul_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f32m4((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwmul_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwmul_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f32m8((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwmul_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwmacc_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f32m2((vfloat32m2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f32m4((vfloat32m4_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f32m8((vfloat32m8_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f32mf2((vfloat32mf2_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f32m1((vfloat32m1_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f32m2((vfloat32m2_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f32m4((vfloat32m4_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f32m8((vfloat32m8_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f32m2((vfloat32m2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f32m4((vfloat32m4_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f32m8((vfloat32m8_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f32mf2((vfloat32mf2_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f32m1((vfloat32m1_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f32m2((vfloat32m2_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f32m4((vfloat32m4_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f32m8((vfloat32m8_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f32m2((vfloat32m2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f32m4((vfloat32m4_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f32m8((vfloat32m8_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f32mf2((vfloat32mf2_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f32m1((vfloat32m1_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f32m2((vfloat32m2_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f32m4((vfloat32m4_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f32m8((vfloat32m8_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f32m2((vfloat32m2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f32m4((vfloat32m4_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f32m8((vfloat32m8_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f32mf2((vfloat32mf2_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f32m1((vfloat32m1_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f32m2((vfloat32m2_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f32m4((vfloat32m4_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f32m8((vfloat32m8_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f16m1_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f16m1_f32m1((vfloat32m1_t)(op0), (vfloat16m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f16m1_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f16m1_f32m1_m((vbool16_t)(op0), (vfloat32m1_t)(op1), (vfloat16m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f16m2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f16m2_f32m1((vfloat32m1_t)(op0), (vfloat16m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f16m2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f16m2_f32m1_m((vbool8_t)(op0), (vfloat32m1_t)(op1), (vfloat16m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f16m4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f16m4_f32m1((vfloat32m1_t)(op0), (vfloat16m4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f16m4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f16m4_f32m1_m((vbool4_t)(op0), (vfloat32m1_t)(op1), (vfloat16m4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f16m8_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f16m8_f32m1((vfloat32m1_t)(op0), (vfloat16m8_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f16m8_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f16m8_f32m1_m((vbool2_t)(op0), (vfloat32m1_t)(op1), (vfloat16m8_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f16mf2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f16mf2_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f16mf2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f16mf2_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f16mf4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f16mf4_f32m1((vfloat32m1_t)(op0), (vfloat16mf4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f16mf4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f16mf4_f32m1_m((vbool64_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f16m1_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f16m1_f32m1((vfloat32m1_t)(op0), (vfloat16m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f16m1_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f16m1_f32m1_m((vbool16_t)(op0), (vfloat32m1_t)(op1), (vfloat16m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f16m2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f16m2_f32m1((vfloat32m1_t)(op0), (vfloat16m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f16m2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f16m2_f32m1_m((vbool8_t)(op0), (vfloat32m1_t)(op1), (vfloat16m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f16m4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f16m4_f32m1((vfloat32m1_t)(op0), (vfloat16m4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f16m4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f16m4_f32m1_m((vbool4_t)(op0), (vfloat32m1_t)(op1), (vfloat16m4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f16m8_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f16m8_f32m1((vfloat32m1_t)(op0), (vfloat16m8_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f16m8_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f16m8_f32m1_m((vbool2_t)(op0), (vfloat32m1_t)(op1), (vfloat16m8_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f16mf2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f16mf2_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f16mf2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f16mf2_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f16mf4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f16mf4_f32m1((vfloat32m1_t)(op0), (vfloat16mf4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f16mf4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f16mf4_f32m1_m((vbool64_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfncvt_f_f_w_f16mf4(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f16mf4((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f16mf2(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f16mf2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f16m1(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f16m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f16m2(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f16m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f16m4(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f16m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f16mf4(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16mf4((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f16mf2(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16mf2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f16m1(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f16m2(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f16m4(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f32mf2((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f32m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f32m1((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f32m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f32m2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f32m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f32m4((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f32m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f32m8((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#endif + +#if defined(__riscv_zvamo) +#define vamoswapei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamominei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamominei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamominei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamominei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamominei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamominei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamominei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamominei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamominei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamominei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamominei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamominei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamominei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamominei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamominei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamominei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamominei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamominei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamominei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamominei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamominei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamominei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamominei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamominei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamominei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamominei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamominei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamominei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamominei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamominei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamominei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamominei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamominei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamominei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamominei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamominei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamominei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamominei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamominei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamominei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamominei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamominei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamominei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamominei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamominei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamominei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamominei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamominei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamominei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamominei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamominei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamominei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamominei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamominei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamominei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamominei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamominei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamominei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamominei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamominei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamominei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamominei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamominei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamominei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamominei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamominei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamominei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamominei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamominei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamominei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#endif + +#if defined(__riscv_f) && defined(__riscv_zvamo) +#define vamoswapei8_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f32m4((float *)(op0), (vuint8m1_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint8m1_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f32m8((float *)(op0), (vuint8m2_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint8m2_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f32m4((float *)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint16m2_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f32m8((float *)(op0), (vuint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint16m4_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f32m4((float *)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f32m8((float *)(op0), (vuint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f32m4((float *)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint64m8_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#endif + +#if defined(__riscv_d) && defined(__riscv_zvamo) +#define vamoswapei8_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f64m4((double *)(op0), (vuint8mf2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint8mf2_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f64m8((double *)(op0), (vuint8m1_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint8m1_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f64m4((double *)(op0), (vuint16m1_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint16m1_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f64m8((double *)(op0), (vuint16m2_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint16m2_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f64m4((double *)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint32m2_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f64m8((double *)(op0), (vuint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint32m4_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f64m4((double *)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f64m8((double *)(op0), (vuint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#endif + +#if defined(__riscv_zvlsseg) +#define vloxseg2ei8_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vsseg3e8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_i8m2((int8_t *)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_i8m2((int8_t *)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsseg2e16_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_i16m2((int16_t *)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_i16m4((int16_t *)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_u16m4((uint16_t *)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_i16m2((int16_t *)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_i16m2((int16_t *)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsseg2e32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_i32m2((int32_t *)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_i32m4((int32_t *)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_i32m2((int32_t *)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_i32m2((int32_t *)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsseg5e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsseg5e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsseg5e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsseg6e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsseg6e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsseg6e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsseg7e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsseg7e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsseg7e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsseg8e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsseg8e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsseg8e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsseg2e64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_i64m2((int64_t *)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_i64m4((int64_t *)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsseg3e64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e64_v_i64m2((int64_t *)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsseg3e64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsseg3e64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsseg4e64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e64_v_i64m2((int64_t *)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsseg4e64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsseg4e64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsseg5e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsseg5e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsseg5e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsseg5e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsseg6e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsseg6e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsseg6e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsseg6e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsseg7e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsseg7e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsseg7e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsseg7e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsseg8e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsseg8e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsseg8e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsseg8e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vssseg3e8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_i8m2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_u8m2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_i8m2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_u8m2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vssseg2e16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_i16m2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_i16m4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_u16m2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_u16m4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_i16m2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_u16m2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_i16m2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_u16m2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vssseg2e32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_i32m2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_i32m4((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_u32m2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_u32m4((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_i32m2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_u32m2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_i32m2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_u32m2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vssseg5e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vssseg5e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vssseg5e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vssseg6e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vssseg6e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vssseg6e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vssseg7e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vssseg7e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vssseg7e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vssseg8e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vssseg8e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vssseg8e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vssseg2e64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_i64m2((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_i64m4((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_u64m2((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_u64m4((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vssseg3e64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e64_v_i64m2((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vssseg3e64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vssseg3e64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e64_v_u64m2((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vssseg4e64_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e64_v_i64m2((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vssseg4e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vssseg4e64_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e64_v_u64m2((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vssseg5e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vssseg5e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vssseg5e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vssseg5e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vssseg6e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vssseg6e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vssseg6e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vssseg6e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vssseg7e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vssseg7e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vssseg7e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vssseg7e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vssseg8e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vssseg8e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vssseg8e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vssseg8e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg3ei8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i8m4((int8_t *)(op0), (vuint16m8_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint16m8_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u8m4((uint8_t *)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i16m4((int16_t *)(op0), (vuint8m2_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint8m2_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u16m4((uint16_t *)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint8m2_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i16m4((int16_t *)(op0), (vuint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint16m4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u16m4((uint16_t *)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i16m4((int16_t *)(op0), (vuint32m8_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint32m8_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u16m4((uint16_t *)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i8m4((int8_t *)(op0), (vuint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint8m4_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u8m4((uint8_t *)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i8m4((int8_t *)(op0), (vuint16m8_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint16m8_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u8m4((uint8_t *)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i16m4((int16_t *)(op0), (vuint8m2_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint8m2_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u16m4((uint16_t *)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint8m2_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i16m4((int16_t *)(op0), (vuint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint16m4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u16m4((uint16_t *)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i16m4((int16_t *)(op0), (vuint32m8_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint32m8_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u16m4((uint16_t *)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vlseg2e8_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg3e8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg4e8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (size_t)(op10)) +#define vlseg5e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (size_t)(op12)) +#define vlseg6e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (size_t)(op14)) +#define vlseg7e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (size_t)(op16)) +#define vlseg8e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (size_t)(op18)) +#define vlseg2e16_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg3e16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg4e16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (size_t)(op10)) +#define vlseg5e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg5e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (size_t)(op12)) +#define vlseg5e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg5e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (size_t)(op12)) +#define vlseg5e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg5e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (size_t)(op12)) +#define vlseg5e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg5e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (size_t)(op12)) +#define vlseg5e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg5e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (size_t)(op12)) +#define vlseg5e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg5e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (size_t)(op12)) +#define vlseg6e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (size_t)(op7)) +#define vlseg6e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (size_t)(op14)) +#define vlseg6e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (size_t)(op7)) +#define vlseg6e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (size_t)(op14)) +#define vlseg6e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (size_t)(op7)) +#define vlseg6e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (size_t)(op14)) +#define vlseg6e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (size_t)(op7)) +#define vlseg6e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (size_t)(op14)) +#define vlseg6e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (size_t)(op7)) +#define vlseg6e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (size_t)(op14)) +#define vlseg6e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (size_t)(op7)) +#define vlseg6e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (size_t)(op14)) +#define vlseg7e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg7e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (size_t)(op16)) +#define vlseg7e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg7e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (size_t)(op16)) +#define vlseg7e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg7e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (size_t)(op16)) +#define vlseg7e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg7e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (size_t)(op16)) +#define vlseg7e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg7e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (size_t)(op16)) +#define vlseg7e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg7e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (size_t)(op16)) +#define vlseg8e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (size_t)(op9)) +#define vlseg8e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (size_t)(op18)) +#define vlseg8e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (size_t)(op9)) +#define vlseg8e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (size_t)(op18)) +#define vlseg8e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (size_t)(op9)) +#define vlseg8e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (size_t)(op18)) +#define vlseg8e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (size_t)(op9)) +#define vlseg8e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (size_t)(op18)) +#define vlseg8e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (size_t)(op9)) +#define vlseg8e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (size_t)(op18)) +#define vlseg8e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (size_t)(op9)) +#define vlseg8e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (size_t)(op18)) +#define vlseg2e32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (size_t)(op6)) +#define vlseg3e32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (size_t)(op4)) +#define vlseg3e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (size_t)(op8)) +#define vlseg3e32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (size_t)(op4)) +#define vlseg3e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (size_t)(op8)) +#define vlseg3e32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (size_t)(op4)) +#define vlseg3e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (size_t)(op8)) +#define vlseg3e32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (size_t)(op4)) +#define vlseg3e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (size_t)(op8)) +#define vlseg3e32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (size_t)(op4)) +#define vlseg3e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (size_t)(op8)) +#define vlseg3e32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (size_t)(op4)) +#define vlseg3e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (size_t)(op8)) +#define vlseg4e32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (size_t)(op5)) +#define vlseg4e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (size_t)(op10)) +#define vlseg4e32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (size_t)(op5)) +#define vlseg4e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (size_t)(op10)) +#define vlseg4e32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (size_t)(op5)) +#define vlseg4e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (size_t)(op10)) +#define vlseg4e32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (size_t)(op5)) +#define vlseg4e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (size_t)(op10)) +#define vlseg4e32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (size_t)(op5)) +#define vlseg4e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (size_t)(op10)) +#define vlseg4e32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (size_t)(op5)) +#define vlseg4e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (size_t)(op10)) +#define vlseg5e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (size_t)(op6)) +#define vlseg5e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (size_t)(op12)) +#define vlseg5e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (size_t)(op6)) +#define vlseg5e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (size_t)(op12)) +#define vlseg5e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (size_t)(op6)) +#define vlseg5e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (size_t)(op12)) +#define vlseg5e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (size_t)(op6)) +#define vlseg5e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (size_t)(op12)) +#define vlseg6e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (size_t)(op7)) +#define vlseg6e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (size_t)(op14)) +#define vlseg6e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (size_t)(op7)) +#define vlseg6e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (size_t)(op14)) +#define vlseg6e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (size_t)(op7)) +#define vlseg6e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (size_t)(op14)) +#define vlseg6e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (size_t)(op7)) +#define vlseg6e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (size_t)(op14)) +#define vlseg7e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (size_t)(op8)) +#define vlseg7e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (size_t)(op16)) +#define vlseg7e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (size_t)(op8)) +#define vlseg7e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (size_t)(op16)) +#define vlseg7e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (size_t)(op8)) +#define vlseg7e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (size_t)(op16)) +#define vlseg7e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (size_t)(op8)) +#define vlseg7e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (size_t)(op16)) +#define vlseg8e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (size_t)(op9)) +#define vlseg8e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (size_t)(op18)) +#define vlseg8e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (size_t)(op9)) +#define vlseg8e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (size_t)(op18)) +#define vlseg8e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (size_t)(op9)) +#define vlseg8e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (size_t)(op18)) +#define vlseg8e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (size_t)(op9)) +#define vlseg8e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (size_t)(op18)) +#define vlseg2e64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vlseg2e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (size_t)(op6)) +#define vlseg2e64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vlseg2e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (size_t)(op6)) +#define vlseg2e64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vlseg2e64_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (size_t)(op6)) +#define vlseg2e64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vlseg2e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (size_t)(op6)) +#define vlseg2e64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vlseg2e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (size_t)(op6)) +#define vlseg2e64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vlseg2e64_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (size_t)(op6)) +#define vlseg3e64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (size_t)(op4)) +#define vlseg3e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (size_t)(op8)) +#define vlseg3e64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (size_t)(op4)) +#define vlseg3e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (size_t)(op8)) +#define vlseg3e64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (size_t)(op4)) +#define vlseg3e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (size_t)(op8)) +#define vlseg3e64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (size_t)(op4)) +#define vlseg3e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (size_t)(op8)) +#define vlseg4e64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (size_t)(op5)) +#define vlseg4e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (size_t)(op10)) +#define vlseg4e64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (size_t)(op5)) +#define vlseg4e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (size_t)(op10)) +#define vlseg4e64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (size_t)(op5)) +#define vlseg4e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (size_t)(op10)) +#define vlseg4e64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (size_t)(op5)) +#define vlseg4e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (size_t)(op10)) +#define vlseg5e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (size_t)(op6)) +#define vlseg5e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (size_t)(op12)) +#define vlseg5e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (size_t)(op6)) +#define vlseg5e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (size_t)(op12)) +#define vlseg6e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (size_t)(op7)) +#define vlseg6e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (size_t)(op14)) +#define vlseg6e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (size_t)(op7)) +#define vlseg6e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (size_t)(op14)) +#define vlseg7e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (size_t)(op8)) +#define vlseg7e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (size_t)(op16)) +#define vlseg7e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (size_t)(op8)) +#define vlseg7e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (size_t)(op16)) +#define vlseg8e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (size_t)(op9)) +#define vlseg8e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (size_t)(op18)) +#define vlseg8e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (size_t)(op9)) +#define vlseg8e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (size_t)(op18)) +#define vlseg3e8ff_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e8ff_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e8ff_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e8ff_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e8ff_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e8ff_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg2e16ff_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg3e16ff_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e16ff_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e16ff_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e16ff_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e16ff_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e16ff_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg2e32ff_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg3e32ff_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e32ff_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e32ff_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e32ff_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e32ff_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e32ff_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e32ff_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e32ff_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e32ff_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e32ff_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e32ff_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e32ff_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg2e8ff_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg7e32ff_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e32ff_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e32ff_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e32ff_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e32ff_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e32ff_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg2e64ff_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg3e64ff_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e64ff_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e64ff_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e64ff_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e64ff_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e64ff_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e64ff_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e64ff_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e64ff_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e64ff_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e64ff_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e64ff_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e64ff_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e64ff_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e64ff_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e64ff_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e64ff_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e64ff_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e64ff_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e64ff_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e64ff_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlsseg2e8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlseg5e64ff_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e64ff_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e64ff_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e64ff_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e64ff_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e64ff_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e64ff_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlsseg2e8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (vuint8m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vlsseg3e8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vluxseg2ei8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (vuint8m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vlsseg3e8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vsseg2e8_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_i8m2((int8_t *)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_i8m4((int8_t *)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vlsseg8e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg2e16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg3e16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vsseg2e8_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_u8m4((uint8_t *)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vlsseg6e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg2e32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg3e32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg2e64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg3e64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e64_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e64_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vluxseg3ei8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (vuint8m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (vuint8m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (vuint8m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (vuint8m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint8m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint8m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint8m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint8m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint8m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint8m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint8m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint8m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (vuint16m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (vuint16m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (vuint16m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (vuint16m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (vuint16m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (vuint16m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (vuint16m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (vuint16m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint16m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint16m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint16m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint16m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint16m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint16m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint16m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint16m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint16m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint16m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint16m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint16m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (vuint32m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (vuint32m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (vuint32m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (vuint32m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (vuint32m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (vuint32m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (vuint32m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (vuint32m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint32m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint32m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint32m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint32m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint32m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint32m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint32m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint32m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint32m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint32m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint32m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint32m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint64m8_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint64m8_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint64m8_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint64m8_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint64m8_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint64m8_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint64m8_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint64m8_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint64m8_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint64m8_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint64m8_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint64m8_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg2ei8_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg3ei8_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vssseg2e8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_i8m2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_i8m4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_i8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg8ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vssseg2e8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_u8m2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_u8m4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_u8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg6ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg2ei8_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg3ei8_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i8m4((int8_t *)(op0), (vuint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint8m4_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg4ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg2ei8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u8m4((uint8_t *)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg2ei16_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg2ei8_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei8_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg2ei8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (vuint8m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (vuint8m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (vuint8m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (vuint8m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (vuint8m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (vuint8m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint8m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint8m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint8m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint8m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint8m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint8m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint8m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint8m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (vuint16m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (vuint16m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (vuint16m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (vuint16m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (vuint16m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (vuint16m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (vuint16m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (vuint16m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint16m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint16m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint16m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint16m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint16m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint16m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint16m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint16m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint16m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint16m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint16m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint16m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (vuint32m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (vuint32m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (vuint32m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (vuint32m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (vuint32m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (vuint32m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (vuint32m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (vuint32m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint32m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint32m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint32m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint32m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint32m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint32m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint32m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint32m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint32m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint32m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint32m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint32m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint64m8_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint64m8_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint64m8_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint64m8_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint64m8_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint64m8_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint64m8_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint64m8_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint64m8_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint64m8_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint64m8_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint64m8_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg2ei8_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg2ei8_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#endif + +#if defined(__riscv_f) && defined(__riscv_zvlsseg) +#define vloxseg2ei8_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vsseg2e32_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_f32m2((float *)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_f32m4((float *)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_f32m2((float *)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_f32m2((float *)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsseg5e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsseg6e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsseg7e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsseg8e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vssseg2e32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_f32m2((float *)(op0), (ptrdiff_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_f32m4((float *)(op0), (ptrdiff_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_f32m2((float *)(op0), (ptrdiff_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_f32m2((float *)(op0), (ptrdiff_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vssseg5e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vssseg6e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vssseg7e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vssseg8e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f32m4((float *)(op0), (vuint8m1_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint8m1_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f32m4((float *)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint16m2_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f32m4((float *)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint32m4_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f32m4((float *)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint64m8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f32m4((float *)(op0), (vuint8m1_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint8m1_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f32m4((float *)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint16m2_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f32m4((float *)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint32m4_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f32m4((float *)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint64m8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vlseg2e32_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (size_t)(op3)) +#define vlseg2e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (size_t)(op6)) +#define vlseg2e32_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (size_t)(op3)) +#define vlseg2e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (size_t)(op6)) +#define vlseg2e32_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (size_t)(op3)) +#define vlseg2e32_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (size_t)(op6)) +#define vlseg2e32_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (size_t)(op3)) +#define vlseg2e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (size_t)(op6)) +#define vlseg3e32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (size_t)(op4)) +#define vlseg3e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (size_t)(op8)) +#define vlseg3e32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (size_t)(op4)) +#define vlseg3e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (size_t)(op8)) +#define vlseg3e32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (size_t)(op4)) +#define vlseg3e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (size_t)(op8)) +#define vlseg4e32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (size_t)(op5)) +#define vlseg4e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (size_t)(op10)) +#define vlseg4e32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (size_t)(op5)) +#define vlseg4e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (size_t)(op10)) +#define vlseg4e32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (size_t)(op5)) +#define vlseg4e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (size_t)(op10)) +#define vlseg5e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (size_t)(op6)) +#define vlseg5e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (size_t)(op12)) +#define vlseg5e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (size_t)(op6)) +#define vlseg5e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (size_t)(op12)) +#define vlseg6e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (size_t)(op7)) +#define vlseg6e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (size_t)(op14)) +#define vlseg6e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (size_t)(op7)) +#define vlseg6e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (size_t)(op14)) +#define vlseg7e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (size_t)(op8)) +#define vlseg7e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (size_t)(op16)) +#define vlseg7e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (size_t)(op8)) +#define vlseg7e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (size_t)(op16)) +#define vlseg8e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (size_t)(op9)) +#define vlseg8e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (size_t)(op18)) +#define vlseg8e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (size_t)(op9)) +#define vlseg8e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (size_t)(op18)) +#define vlseg2e32ff_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg3e32ff_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e32ff_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e32ff_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e32ff_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e32ff_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e32ff_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e32ff_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e32ff_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e32ff_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e32ff_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlsseg2e32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg3e32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vluxseg2ei8_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg3ei8_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#endif + +#if defined(__riscv_d) && defined(__riscv_zvlsseg) +#define vloxseg2ei8_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vsseg2e64_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_f64m2((double *)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_f64m4((double *)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsseg3e64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e64_v_f64m2((double *)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsseg4e64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e64_v_f64m2((double *)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsseg5e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsseg5e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsseg6e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsseg6e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsseg7e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsseg7e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsseg8e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsseg8e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vssseg2e64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_f64m2((double *)(op0), (ptrdiff_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_f64m4((double *)(op0), (ptrdiff_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vssseg3e64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e64_v_f64m2((double *)(op0), (ptrdiff_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vssseg4e64_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e64_v_f64m2((double *)(op0), (ptrdiff_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vssseg5e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vssseg5e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vssseg6e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vssseg6e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vssseg7e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vssseg7e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vssseg8e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vssseg8e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f64m4((double *)(op0), (vuint8mf2_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint8mf2_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f64m4((double *)(op0), (vuint16m1_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint16m1_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f64m4((double *)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint32m2_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f64m4((double *)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint64m4_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f64m4((double *)(op0), (vuint8mf2_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint8mf2_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f64m4((double *)(op0), (vuint16m1_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint16m1_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f64m4((double *)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint32m2_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f64m4((double *)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint64m4_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vlseg2e64_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (size_t)(op3)) +#define vlseg2e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (size_t)(op6)) +#define vlseg2e64_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (size_t)(op3)) +#define vlseg2e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (size_t)(op6)) +#define vlseg2e64_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (size_t)(op3)) +#define vlseg2e64_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (size_t)(op6)) +#define vlseg3e64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (size_t)(op4)) +#define vlseg3e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (size_t)(op8)) +#define vlseg3e64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (size_t)(op4)) +#define vlseg3e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (size_t)(op8)) +#define vlseg4e64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (size_t)(op5)) +#define vlseg4e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (size_t)(op10)) +#define vlseg4e64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (size_t)(op5)) +#define vlseg4e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (size_t)(op10)) +#define vlseg5e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (size_t)(op6)) +#define vlseg5e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (size_t)(op12)) +#define vlseg6e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (size_t)(op7)) +#define vlseg6e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (size_t)(op14)) +#define vlseg7e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (size_t)(op8)) +#define vlseg7e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (size_t)(op16)) +#define vlseg8e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (size_t)(op9)) +#define vlseg8e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (size_t)(op18)) +#define vlseg2e64ff_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg3e64ff_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e64ff_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e64ff_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e64ff_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e64ff_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e64ff_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e64ff_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e64ff_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e64ff_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e64ff_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e64ff_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e64ff_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e64ff_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e64ff_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlsseg2e64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg3e64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e64_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vluxseg2ei8_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei8_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#endif + +#if defined(__riscv_zfh) && defined(__riscv_zvlsseg) +#define vloxseg2ei8_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vsseg2e16_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_f16m2((_Float16 *)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_f16m4((_Float16 *)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_f16m2((_Float16 *)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_f16m2((_Float16 *)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vssseg2e16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_f16m2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_f16m4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_f16m2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_f16m2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f16m4((_Float16 *)(op0), (vuint8m2_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint8m2_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f16m4((_Float16 *)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint16m4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f16m4((_Float16 *)(op0), (vuint32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint32m8_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f16m4((_Float16 *)(op0), (vuint8m2_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint8m2_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f16m4((_Float16 *)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint16m4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f16m4((_Float16 *)(op0), (vuint32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint32m8_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vlseg2e16_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vlseg2e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg2e16_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vlseg2e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg2e16_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vlseg2e16_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg2e16_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vlseg2e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg2e16_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vlseg2e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg3e16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (size_t)(op4)) +#define vlseg3e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg3e16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (size_t)(op4)) +#define vlseg3e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg3e16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (size_t)(op4)) +#define vlseg3e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg3e16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (size_t)(op4)) +#define vlseg3e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg4e16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (size_t)(op5)) +#define vlseg4e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (size_t)(op10)) +#define vlseg4e16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (size_t)(op5)) +#define vlseg4e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (size_t)(op10)) +#define vlseg4e16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (size_t)(op5)) +#define vlseg4e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (size_t)(op10)) +#define vlseg4e16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (size_t)(op5)) +#define vlseg4e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (size_t)(op10)) +#define vlseg5e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg5e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (size_t)(op12)) +#define vlseg5e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg5e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (size_t)(op12)) +#define vlseg5e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg5e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (size_t)(op12)) +#define vlseg6e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (size_t)(op7)) +#define vlseg6e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (size_t)(op14)) +#define vlseg6e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (size_t)(op7)) +#define vlseg6e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (size_t)(op14)) +#define vlseg6e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (size_t)(op7)) +#define vlseg6e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (size_t)(op14)) +#define vlseg7e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg7e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (size_t)(op16)) +#define vlseg7e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg7e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (size_t)(op16)) +#define vlseg7e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg7e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (size_t)(op16)) +#define vlseg8e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (size_t)(op9)) +#define vlseg8e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (size_t)(op18)) +#define vlseg8e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (size_t)(op9)) +#define vlseg8e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (size_t)(op18)) +#define vlseg8e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (size_t)(op9)) +#define vlseg8e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (size_t)(op18)) +#define vlseg2e16ff_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg3e16ff_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e16ff_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e16ff_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e16ff_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e16ff_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e16ff_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlsseg2e16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg3e16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vluxseg2ei8_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg3ei8_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#endif + +#define __riscv_v_intrinsic_overloading 1 +#define __rvv_overloaded static inline __attribute__((__always_inline__, __nodebug__, __overloadable__)) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m1))) +vint8m1_t vadd(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m1_m))) +vint8m1_t vadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m2))) +vint8m2_t vadd(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m2_m))) +vint8m2_t vadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m4))) +vint8m4_t vadd(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m4_m))) +vint8m4_t vadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m8))) +vint8m8_t vadd(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m8_m))) +vint8m8_t vadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8mf2))) +vint8mf2_t vadd(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8mf2_m))) +vint8mf2_t vadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8mf4))) +vint8mf4_t vadd(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8mf4_m))) +vint8mf4_t vadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8mf8))) +vint8mf8_t vadd(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8mf8_m))) +vint8mf8_t vadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m1))) +vint16m1_t vadd(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m1_m))) +vint16m1_t vadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m2))) +vint16m2_t vadd(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m2_m))) +vint16m2_t vadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m4))) +vint16m4_t vadd(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m4_m))) +vint16m4_t vadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m8))) +vint16m8_t vadd(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m8_m))) +vint16m8_t vadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16mf2))) +vint16mf2_t vadd(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16mf2_m))) +vint16mf2_t vadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16mf4))) +vint16mf4_t vadd(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16mf4_m))) +vint16mf4_t vadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m1))) +vint32m1_t vadd(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m1_m))) +vint32m1_t vadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m2))) +vint32m2_t vadd(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m2_m))) +vint32m2_t vadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m4))) +vint32m4_t vadd(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m4_m))) +vint32m4_t vadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m8))) +vint32m8_t vadd(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m8_m))) +vint32m8_t vadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32mf2))) +vint32mf2_t vadd(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32mf2_m))) +vint32mf2_t vadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m1))) +vint64m1_t vadd(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m1_m))) +vint64m1_t vadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m2))) +vint64m2_t vadd(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m2_m))) +vint64m2_t vadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m4))) +vint64m4_t vadd(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m4_m))) +vint64m4_t vadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m8))) +vint64m8_t vadd(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m8_m))) +vint64m8_t vadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16mf4))) +vuint16mf4_t vwaddu_vv(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16mf4_m))) +vuint16mf4_t vwaddu_vv(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16mf2))) +vuint16mf2_t vwaddu_vv(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16mf2_m))) +vuint16mf2_t vwaddu_vv(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m1))) +vuint16m1_t vwaddu_vv(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m1_m))) +vuint16m1_t vwaddu_vv(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m2))) +vuint16m2_t vwaddu_vv(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m2_m))) +vuint16m2_t vwaddu_vv(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m4))) +vuint16m4_t vwaddu_vv(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m4_m))) +vuint16m4_t vwaddu_vv(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m8))) +vuint16m8_t vwaddu_vv(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m8_m))) +vuint16m8_t vwaddu_vv(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32mf2))) +vuint32mf2_t vwaddu_vv(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32mf2_m))) +vuint32mf2_t vwaddu_vv(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m1))) +vuint32m1_t vwaddu_vv(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m1_m))) +vuint32m1_t vwaddu_vv(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m2))) +vuint32m2_t vwaddu_vv(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m2_m))) +vuint32m2_t vwaddu_vv(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m4))) +vuint32m4_t vwaddu_vv(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m4_m))) +vuint32m4_t vwaddu_vv(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m8))) +vuint32m8_t vwaddu_vv(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m8_m))) +vuint32m8_t vwaddu_vv(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m1))) +vuint64m1_t vwaddu_vv(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m1_m))) +vuint64m1_t vwaddu_vv(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m2))) +vuint64m2_t vwaddu_vv(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m2_m))) +vuint64m2_t vwaddu_vv(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m4))) +vuint64m4_t vwaddu_vv(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m4_m))) +vuint64m4_t vwaddu_vv(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m8))) +vuint64m8_t vwaddu_vv(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m8_m))) +vuint64m8_t vwaddu_vv(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m1))) +vuint8m1_t vluxei8(const uint8_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m1_m))) +vuint8m1_t vluxei8(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m2))) +vuint8m2_t vluxei8(const uint8_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m2_m))) +vuint8m2_t vluxei8(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m4))) +vuint8m4_t vluxei8(const uint8_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m4_m))) +vuint8m4_t vluxei8(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m8))) +vuint8m8_t vluxei8(const uint8_t * op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m8_m))) +vuint8m8_t vluxei8(vbool1_t op0, vuint8m8_t op1, const uint8_t * op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8mf2))) +vuint8mf2_t vluxei8(const uint8_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8mf2_m))) +vuint8mf2_t vluxei8(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8mf4))) +vuint8mf4_t vluxei8(const uint8_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8mf4_m))) +vuint8mf4_t vluxei8(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8mf8))) +vuint8mf8_t vluxei8(const uint8_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8mf8_m))) +vuint8mf8_t vluxei8(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_u64m1_m))) +vuint64m1_t vlse64(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_u64m2_m))) +vuint64m2_t vlse64(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_u64m4_m))) +vuint64m4_t vlse64(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_u64m8_m))) +vuint64m8_t vlse64(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m1))) +void vsse16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m1_m))) +void vsse16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m2))) +void vsse16(int16_t * op0, ptrdiff_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m2_m))) +void vsse16(vbool8_t op0, int16_t * op1, ptrdiff_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m4))) +void vsse16(int16_t * op0, ptrdiff_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m4_m))) +void vsse16(vbool4_t op0, int16_t * op1, ptrdiff_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m8))) +void vsse16(int16_t * op0, ptrdiff_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m8_m))) +void vsse16(vbool2_t op0, int16_t * op1, ptrdiff_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16mf2))) +void vsse16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16mf2_m))) +void vsse16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16mf4))) +void vsse16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16mf4_m))) +void vsse16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m1))) +void vsse16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m1_m))) +void vsse16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m2))) +void vsse16(uint16_t * op0, ptrdiff_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m2_m))) +void vsse16(vbool8_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m4))) +void vsse16(uint16_t * op0, ptrdiff_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m4_m))) +void vsse16(vbool4_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m8))) +void vsse16(uint16_t * op0, ptrdiff_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m8_m))) +void vsse16(vbool2_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16mf2))) +void vsse16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16mf2_m))) +void vsse16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16mf4))) +void vsse16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16mf4_m))) +void vsse16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m1))) +void vsse32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m1_m))) +void vsse32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m2))) +void vsse32(int32_t * op0, ptrdiff_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m2_m))) +void vsse32(vbool16_t op0, int32_t * op1, ptrdiff_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m4))) +void vsse32(int32_t * op0, ptrdiff_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m4_m))) +void vsse32(vbool8_t op0, int32_t * op1, ptrdiff_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m8))) +void vsse32(int32_t * op0, ptrdiff_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m8_m))) +void vsse32(vbool4_t op0, int32_t * op1, ptrdiff_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32mf2))) +void vsse32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32mf2_m))) +void vsse32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m1))) +void vsse32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m1_m))) +void vsse32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m2))) +void vsse32(uint32_t * op0, ptrdiff_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m2_m))) +void vsse32(vbool16_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m4))) +void vsse32(uint32_t * op0, ptrdiff_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m4_m))) +void vsse32(vbool8_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m8))) +void vsse32(uint32_t * op0, ptrdiff_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m8_m))) +void vsse32(vbool4_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32mf2))) +void vsse32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32mf2_m))) +void vsse32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m1))) +void vsse64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m1_m))) +void vsse64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m2))) +void vsse64(int64_t * op0, ptrdiff_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m2_m))) +void vsse64(vbool32_t op0, int64_t * op1, ptrdiff_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m4))) +void vsse64(int64_t * op0, ptrdiff_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m4_m))) +void vsse64(vbool16_t op0, int64_t * op1, ptrdiff_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m8))) +void vsse64(int64_t * op0, ptrdiff_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m8_m))) +void vsse64(vbool8_t op0, int64_t * op1, ptrdiff_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m1))) +void vsse64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m1_m))) +void vsse64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m2))) +void vsse64(uint64_t * op0, ptrdiff_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m2_m))) +void vsse64(vbool32_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m4))) +void vsse64(uint64_t * op0, ptrdiff_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m4_m))) +void vsse64(vbool16_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m8))) +void vsse64(uint64_t * op0, ptrdiff_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m8_m))) +void vsse64(vbool8_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8m1))) +vint8m1_t vluxei16(const int8_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8m1_m))) +vint8m1_t vluxei16(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8m2))) +vint8m2_t vluxei16(const int8_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8m2_m))) +vint8m2_t vluxei16(vbool4_t op0, vint8m2_t op1, const int8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8m4))) +vint8m4_t vluxei16(const int8_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8m4_m))) +vint8m4_t vluxei16(vbool2_t op0, vint8m4_t op1, const int8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8mf2))) +vint8mf2_t vluxei16(const int8_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8mf2_m))) +vint8mf2_t vluxei16(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8mf4))) +vint8mf4_t vluxei16(const int8_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8mf4_m))) +vint8mf4_t vluxei16(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8mf8))) +vint8mf8_t vluxei16(const int8_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8mf8_m))) +vint8mf8_t vluxei16(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8m1))) +vuint8m1_t vluxei16(const uint8_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8m1_m))) +vuint8m1_t vluxei16(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8m2))) +vuint8m2_t vluxei16(const uint8_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8m2_m))) +vuint8m2_t vluxei16(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8m4))) +vuint8m4_t vluxei16(const uint8_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8m4_m))) +vuint8m4_t vluxei16(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8mf2))) +vuint8mf2_t vluxei16(const uint8_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8mf2_m))) +vuint8mf2_t vluxei16(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8mf4))) +vuint8mf4_t vluxei16(const uint8_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8mf4_m))) +vuint8mf4_t vluxei16(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8mf8))) +vuint8mf8_t vluxei16(const uint8_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8mf8_m))) +vuint8mf8_t vluxei16(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8m1))) +vint8m1_t vluxei32(const int8_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8m1_m))) +vint8m1_t vluxei32(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8m2))) +vint8m2_t vluxei32(const int8_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8m2_m))) +vint8m2_t vluxei32(vbool4_t op0, vint8m2_t op1, const int8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8mf2))) +vint8mf2_t vluxei32(const int8_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8mf2_m))) +vint8mf2_t vluxei32(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8mf4))) +vint8mf4_t vluxei32(const int8_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8mf4_m))) +vint8mf4_t vluxei32(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8mf8))) +vint8mf8_t vluxei32(const int8_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8mf8_m))) +vint8mf8_t vluxei32(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8m1))) +vuint8m1_t vluxei32(const uint8_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8m1_m))) +vuint8m1_t vluxei32(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8m2))) +vuint8m2_t vluxei32(const uint8_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8m2_m))) +vuint8m2_t vluxei32(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8mf2))) +vuint8mf2_t vluxei32(const uint8_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8mf2_m))) +vuint8mf2_t vluxei32(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8mf4))) +vuint8mf4_t vluxei32(const uint8_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8mf4_m))) +vuint8mf4_t vluxei32(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8mf8))) +vuint8mf8_t vluxei32(const uint8_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8mf8_m))) +vuint8mf8_t vluxei32(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8m1))) +vint8m1_t vluxei64(const int8_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8m1_m))) +vint8m1_t vluxei64(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8mf2))) +vint8mf2_t vluxei64(const int8_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8mf2_m))) +vint8mf2_t vluxei64(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8mf4))) +vint8mf4_t vluxei64(const int8_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8mf4_m))) +vint8mf4_t vluxei64(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8mf8))) +vint8mf8_t vluxei64(const int8_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8mf8_m))) +vint8mf8_t vluxei64(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8m1))) +vuint8m1_t vluxei64(const uint8_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8m1_m))) +vuint8m1_t vluxei64(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8mf2))) +vuint8mf2_t vluxei64(const uint8_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8mf2_m))) +vuint8mf2_t vluxei64(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8mf4))) +vuint8mf4_t vluxei64(const uint8_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8mf4_m))) +vuint8mf4_t vluxei64(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8mf8))) +vuint8mf8_t vluxei64(const uint8_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8mf8_m))) +vuint8mf8_t vluxei64(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m1))) +vint16m1_t vluxei8(const int16_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m1_m))) +vint16m1_t vluxei8(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m2))) +vint16m2_t vluxei8(const int16_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m2_m))) +vint16m2_t vluxei8(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m4))) +vint16m4_t vluxei8(const int16_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m4_m))) +vint16m4_t vluxei8(vbool4_t op0, vint16m4_t op1, const int16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m8))) +vint16m8_t vluxei8(const int16_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m8_m))) +vint16m8_t vluxei8(vbool2_t op0, vint16m8_t op1, const int16_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16mf2))) +vint16mf2_t vluxei8(const int16_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16mf2_m))) +vint16mf2_t vluxei8(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16mf4))) +vint16mf4_t vluxei8(const int16_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16mf4_m))) +vint16mf4_t vluxei8(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m1))) +vuint16m1_t vluxei8(const uint16_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m1_m))) +vuint16m1_t vluxei8(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m2))) +vuint16m2_t vluxei8(const uint16_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m2_m))) +vuint16m2_t vluxei8(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m4))) +vuint16m4_t vluxei8(const uint16_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m4_m))) +vuint16m4_t vluxei8(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m8))) +vuint16m8_t vluxei8(const uint16_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m8_m))) +vuint16m8_t vluxei8(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16mf2))) +vuint16mf2_t vluxei8(const uint16_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16mf2_m))) +vuint16mf2_t vluxei8(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16mf4))) +vuint16mf4_t vluxei8(const uint16_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16mf4_m))) +vuint16mf4_t vluxei8(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m1))) +vint16m1_t vluxei16(const int16_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m1_m))) +vint16m1_t vluxei16(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m2))) +vint16m2_t vluxei16(const int16_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m2_m))) +vint16m2_t vluxei16(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m4))) +vint16m4_t vluxei16(const int16_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m4_m))) +vint16m4_t vluxei16(vbool4_t op0, vint16m4_t op1, const int16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m8))) +vint16m8_t vluxei16(const int16_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m8_m))) +vint16m8_t vluxei16(vbool2_t op0, vint16m8_t op1, const int16_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16mf2))) +vint16mf2_t vluxei16(const int16_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16mf2_m))) +vint16mf2_t vluxei16(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16mf4))) +vint16mf4_t vluxei16(const int16_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16mf4_m))) +vint16mf4_t vluxei16(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m1))) +vuint16m1_t vluxei16(const uint16_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m1_m))) +vuint16m1_t vluxei16(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m2))) +vuint16m2_t vluxei16(const uint16_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m2_m))) +vuint16m2_t vluxei16(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m4))) +vuint16m4_t vluxei16(const uint16_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m4_m))) +vuint16m4_t vluxei16(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m8))) +vuint16m8_t vluxei16(const uint16_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m8_m))) +vuint16m8_t vluxei16(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16mf2))) +vuint16mf2_t vluxei16(const uint16_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16mf2_m))) +vuint16mf2_t vluxei16(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16mf4))) +vuint16mf4_t vluxei16(const uint16_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16mf4_m))) +vuint16mf4_t vluxei16(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16m1))) +vint16m1_t vluxei32(const int16_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16m1_m))) +vint16m1_t vluxei32(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16m2))) +vint16m2_t vluxei32(const int16_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16m2_m))) +vint16m2_t vluxei32(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16m4))) +vint16m4_t vluxei32(const int16_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16m4_m))) +vint16m4_t vluxei32(vbool4_t op0, vint16m4_t op1, const int16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16mf2))) +vint16mf2_t vluxei32(const int16_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16mf2_m))) +vint16mf2_t vluxei32(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16mf4))) +vint16mf4_t vluxei32(const int16_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16mf4_m))) +vint16mf4_t vluxei32(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16m1))) +vuint16m1_t vluxei32(const uint16_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16m1_m))) +vuint16m1_t vluxei32(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16m2))) +vuint16m2_t vluxei32(const uint16_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16m2_m))) +vuint16m2_t vluxei32(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16m4))) +vuint16m4_t vluxei32(const uint16_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16m4_m))) +vuint16m4_t vluxei32(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16mf2))) +vuint16mf2_t vluxei32(const uint16_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16mf2_m))) +vuint16mf2_t vluxei32(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16mf4))) +vuint16mf4_t vluxei32(const uint16_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16mf4_m))) +vuint16mf4_t vluxei32(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16m1))) +vint16m1_t vluxei64(const int16_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16m1_m))) +vint16m1_t vluxei64(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16m2))) +vint16m2_t vluxei64(const int16_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16m2_m))) +vint16m2_t vluxei64(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16mf2))) +vint16mf2_t vluxei64(const int16_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16mf2_m))) +vint16mf2_t vluxei64(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16mf4))) +vint16mf4_t vluxei64(const int16_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16mf4_m))) +vint16mf4_t vluxei64(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16m1))) +vuint16m1_t vluxei64(const uint16_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16m1_m))) +vuint16m1_t vluxei64(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16m2))) +vuint16m2_t vluxei64(const uint16_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16m2_m))) +vuint16m2_t vluxei64(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16mf2))) +vuint16mf2_t vluxei64(const uint16_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16mf2_m))) +vuint16mf2_t vluxei64(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16mf4))) +vuint16mf4_t vluxei64(const uint16_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16mf4_m))) +vuint16mf4_t vluxei64(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m1))) +vint32m1_t vluxei8(const int32_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m1_m))) +vint32m1_t vluxei8(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m2))) +vint32m2_t vluxei8(const int32_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m2_m))) +vint32m2_t vluxei8(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m4))) +vint32m4_t vluxei8(const int32_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m4_m))) +vint32m4_t vluxei8(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m8))) +vint32m8_t vluxei8(const int32_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m8_m))) +vint32m8_t vluxei8(vbool4_t op0, vint32m8_t op1, const int32_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32mf2))) +vint32mf2_t vluxei8(const int32_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32mf2_m))) +vint32mf2_t vluxei8(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m1))) +vuint32m1_t vluxei8(const uint32_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m1_m))) +vuint32m1_t vluxei8(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m2))) +vuint32m2_t vluxei8(const uint32_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m2_m))) +vuint32m2_t vluxei8(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m4))) +vuint32m4_t vluxei8(const uint32_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m4_m))) +vuint32m4_t vluxei8(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m8))) +vuint32m8_t vluxei8(const uint32_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m8_m))) +vuint32m8_t vluxei8(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32mf2))) +vuint32mf2_t vluxei8(const uint32_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32mf2_m))) +vuint32mf2_t vluxei8(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m1))) +vint32m1_t vluxei16(const int32_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m1_m))) +vint32m1_t vluxei16(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m2))) +vint32m2_t vluxei16(const int32_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m2_m))) +vint32m2_t vluxei16(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m4))) +vint32m4_t vluxei16(const int32_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m4_m))) +vint32m4_t vluxei16(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m8))) +vint32m8_t vluxei16(const int32_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m8_m))) +vint32m8_t vluxei16(vbool4_t op0, vint32m8_t op1, const int32_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32mf2))) +vint32mf2_t vluxei16(const int32_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32mf2_m))) +vint32mf2_t vluxei16(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m1))) +vuint32m1_t vluxei16(const uint32_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m1_m))) +vuint32m1_t vluxei16(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m2))) +vuint32m2_t vluxei16(const uint32_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m2_m))) +vuint32m2_t vluxei16(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m4))) +vuint32m4_t vluxei16(const uint32_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m4_m))) +vuint32m4_t vluxei16(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m8))) +vuint32m8_t vluxei16(const uint32_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m8_m))) +vuint32m8_t vluxei16(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32mf2))) +vuint32mf2_t vluxei16(const uint32_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32mf2_m))) +vuint32mf2_t vluxei16(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m1))) +vint32m1_t vluxei32(const int32_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m1_m))) +vint32m1_t vluxei32(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m2))) +vint32m2_t vluxei32(const int32_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m2_m))) +vint32m2_t vluxei32(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m4))) +vint32m4_t vluxei32(const int32_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m4_m))) +vint32m4_t vluxei32(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m8))) +vint32m8_t vluxei32(const int32_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m8_m))) +vint32m8_t vluxei32(vbool4_t op0, vint32m8_t op1, const int32_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32mf2))) +vint32mf2_t vluxei32(const int32_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32mf2_m))) +vint32mf2_t vluxei32(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m1))) +vuint32m1_t vluxei32(const uint32_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m1_m))) +vuint32m1_t vluxei32(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m2))) +vuint32m2_t vluxei32(const uint32_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m2_m))) +vuint32m2_t vluxei32(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m4))) +vuint32m4_t vluxei32(const uint32_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m4_m))) +vuint32m4_t vluxei32(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m8))) +vuint32m8_t vluxei32(const uint32_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m8_m))) +vuint32m8_t vluxei32(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32mf2))) +vuint32mf2_t vluxei32(const uint32_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32mf2_m))) +vuint32mf2_t vluxei32(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32m1))) +vint32m1_t vluxei64(const int32_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32m1_m))) +vint32m1_t vluxei64(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32m2))) +vint32m2_t vluxei64(const int32_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32m2_m))) +vint32m2_t vluxei64(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32m4))) +vint32m4_t vluxei64(const int32_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32m4_m))) +vint32m4_t vluxei64(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32mf2))) +vint32mf2_t vluxei64(const int32_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32mf2_m))) +vint32mf2_t vluxei64(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32m1))) +vuint32m1_t vluxei64(const uint32_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32m1_m))) +vuint32m1_t vluxei64(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32m2))) +vuint32m2_t vluxei64(const uint32_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32m2_m))) +vuint32m2_t vluxei64(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32m4))) +vuint32m4_t vluxei64(const uint32_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32m4_m))) +vuint32m4_t vluxei64(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32mf2))) +vuint32mf2_t vluxei64(const uint32_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32mf2_m))) +vuint32mf2_t vluxei64(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m1))) +vint64m1_t vluxei8(const int64_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m1_m))) +vint64m1_t vluxei8(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m2))) +vint64m2_t vluxei8(const int64_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m2_m))) +vint64m2_t vluxei8(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m4))) +vint64m4_t vluxei8(const int64_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m4_m))) +vint64m4_t vluxei8(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m8))) +vint64m8_t vluxei8(const int64_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m8_m))) +vint64m8_t vluxei8(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m1))) +vuint64m1_t vluxei8(const uint64_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m1_m))) +vuint64m1_t vluxei8(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m2))) +vuint64m2_t vluxei8(const uint64_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m2_m))) +vuint64m2_t vluxei8(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m4))) +vuint64m4_t vluxei8(const uint64_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m4_m))) +vuint64m4_t vluxei8(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m8))) +vuint64m8_t vluxei8(const uint64_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m8_m))) +vuint64m8_t vluxei8(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m1))) +vint64m1_t vluxei16(const int64_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m1_m))) +vint64m1_t vluxei16(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m2))) +vint64m2_t vluxei16(const int64_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m2_m))) +vint64m2_t vluxei16(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m4))) +vint64m4_t vluxei16(const int64_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m4_m))) +vint64m4_t vluxei16(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m8))) +vint64m8_t vluxei16(const int64_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m8_m))) +vint64m8_t vluxei16(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m1))) +vuint64m1_t vluxei16(const uint64_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m1_m))) +vuint64m1_t vluxei16(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m2))) +vuint64m2_t vluxei16(const uint64_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m2_m))) +vuint64m2_t vluxei16(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m4))) +vuint64m4_t vluxei16(const uint64_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m4_m))) +vuint64m4_t vluxei16(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m8))) +vuint64m8_t vluxei16(const uint64_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m8_m))) +vuint64m8_t vluxei16(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m1))) +vint64m1_t vluxei32(const int64_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m1_m))) +vint64m1_t vluxei32(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m2))) +vint64m2_t vluxei32(const int64_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m2_m))) +vint64m2_t vluxei32(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m4))) +vint64m4_t vluxei32(const int64_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m4_m))) +vint64m4_t vluxei32(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m8))) +vint64m8_t vluxei32(const int64_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m8_m))) +vint64m8_t vluxei32(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m1))) +vuint64m1_t vluxei32(const uint64_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m1_m))) +vuint64m1_t vluxei32(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m2))) +vuint64m2_t vluxei32(const uint64_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m2_m))) +vuint64m2_t vluxei32(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m4))) +vuint64m4_t vluxei32(const uint64_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m4_m))) +vuint64m4_t vluxei32(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m8))) +vuint64m8_t vluxei32(const uint64_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m8_m))) +vuint64m8_t vluxei32(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m1))) +vint64m1_t vluxei64(const int64_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m1_m))) +vint64m1_t vluxei64(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m2))) +vint64m2_t vluxei64(const int64_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m2_m))) +vint64m2_t vluxei64(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m4))) +vint64m4_t vluxei64(const int64_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m4_m))) +vint64m4_t vluxei64(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m8))) +vint64m8_t vluxei64(const int64_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m8_m))) +vint64m8_t vluxei64(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m1))) +vuint64m1_t vluxei64(const uint64_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m1_m))) +vuint64m1_t vluxei64(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m2))) +vuint64m2_t vluxei64(const uint64_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m2_m))) +vuint64m2_t vluxei64(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m4))) +vuint64m4_t vluxei64(const uint64_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m4_m))) +vuint64m4_t vluxei64(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m8))) +vuint64m8_t vluxei64(const uint64_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m8_m))) +vuint64m8_t vluxei64(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m1))) +vint8m1_t vloxei8(const int8_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m1_m))) +vint8m1_t vloxei8(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m2))) +vint8m2_t vloxei8(const int8_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m2_m))) +vint8m2_t vloxei8(vbool4_t op0, vint8m2_t op1, const int8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m4))) +vint8m4_t vloxei8(const int8_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m4_m))) +vint8m4_t vloxei8(vbool2_t op0, vint8m4_t op1, const int8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m8))) +vint8m8_t vloxei8(const int8_t * op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m8_m))) +vint8m8_t vloxei8(vbool1_t op0, vint8m8_t op1, const int8_t * op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8mf2))) +vint8mf2_t vloxei8(const int8_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8mf2_m))) +vint8mf2_t vloxei8(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8mf4))) +vint8mf4_t vloxei8(const int8_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8mf4_m))) +vint8mf4_t vloxei8(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8mf8))) +vint8mf8_t vloxei8(const int8_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8mf8_m))) +vint8mf8_t vloxei8(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m1))) +vuint8m1_t vloxei8(const uint8_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m1_m))) +vuint8m1_t vloxei8(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m2))) +vuint8m2_t vloxei8(const uint8_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m2_m))) +vuint8m2_t vloxei8(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m4))) +vuint8m4_t vloxei8(const uint8_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m4_m))) +vuint8m4_t vloxei8(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m8))) +vuint8m8_t vloxei8(const uint8_t * op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m8_m))) +vuint8m8_t vloxei8(vbool1_t op0, vuint8m8_t op1, const uint8_t * op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8mf2))) +vuint8mf2_t vloxei8(const uint8_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8mf2_m))) +vuint8mf2_t vloxei8(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8mf4))) +vuint8mf4_t vloxei8(const uint8_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8mf4_m))) +vuint8mf4_t vloxei8(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8mf8))) +vuint8mf8_t vloxei8(const uint8_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8mf8_m))) +vuint8mf8_t vloxei8(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8m1))) +vint8m1_t vloxei16(const int8_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8m1_m))) +vint8m1_t vloxei16(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8m2))) +vint8m2_t vloxei16(const int8_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8m2_m))) +vint8m2_t vloxei16(vbool4_t op0, vint8m2_t op1, const int8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8m4))) +vint8m4_t vloxei16(const int8_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8m4_m))) +vint8m4_t vloxei16(vbool2_t op0, vint8m4_t op1, const int8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8mf2))) +vint8mf2_t vloxei16(const int8_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8mf2_m))) +vint8mf2_t vloxei16(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8mf4))) +vint8mf4_t vloxei16(const int8_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8mf4_m))) +vint8mf4_t vloxei16(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8mf8))) +vint8mf8_t vloxei16(const int8_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8mf8_m))) +vint8mf8_t vloxei16(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8m1))) +vuint8m1_t vloxei16(const uint8_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8m1_m))) +vuint8m1_t vloxei16(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8m2))) +vuint8m2_t vloxei16(const uint8_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8m2_m))) +vuint8m2_t vloxei16(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8m4))) +vuint8m4_t vloxei16(const uint8_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8m4_m))) +vuint8m4_t vloxei16(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8mf2))) +vuint8mf2_t vloxei16(const uint8_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8mf2_m))) +vuint8mf2_t vloxei16(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8mf4))) +vuint8mf4_t vloxei16(const uint8_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8mf4_m))) +vuint8mf4_t vloxei16(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8mf8))) +vuint8mf8_t vloxei16(const uint8_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8mf8_m))) +vuint8mf8_t vloxei16(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8m1))) +vint8m1_t vloxei32(const int8_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8m1_m))) +vint8m1_t vloxei32(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8m2))) +vint8m2_t vloxei32(const int8_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8m2_m))) +vint8m2_t vloxei32(vbool4_t op0, vint8m2_t op1, const int8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8mf2))) +vint8mf2_t vloxei32(const int8_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8mf2_m))) +vint8mf2_t vloxei32(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8mf4))) +vint8mf4_t vloxei32(const int8_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8mf4_m))) +vint8mf4_t vloxei32(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8mf8))) +vint8mf8_t vloxei32(const int8_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8mf8_m))) +vint8mf8_t vloxei32(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8m1))) +vuint8m1_t vloxei32(const uint8_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8m1_m))) +vuint8m1_t vloxei32(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8m2))) +vuint8m2_t vloxei32(const uint8_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8m2_m))) +vuint8m2_t vloxei32(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8mf2))) +vuint8mf2_t vloxei32(const uint8_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8mf2_m))) +vuint8mf2_t vloxei32(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8mf4))) +vuint8mf4_t vloxei32(const uint8_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8mf4_m))) +vuint8mf4_t vloxei32(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8mf8))) +vuint8mf8_t vloxei32(const uint8_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8mf8_m))) +vuint8mf8_t vloxei32(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8m1))) +vint8m1_t vloxei64(const int8_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8m1_m))) +vint8m1_t vloxei64(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8mf2))) +vint8mf2_t vloxei64(const int8_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8mf2_m))) +vint8mf2_t vloxei64(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8mf4))) +vint8mf4_t vloxei64(const int8_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8mf4_m))) +vint8mf4_t vloxei64(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8mf8))) +vint8mf8_t vloxei64(const int8_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8mf8_m))) +vint8mf8_t vloxei64(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8m1))) +vuint8m1_t vloxei64(const uint8_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8m1_m))) +vuint8m1_t vloxei64(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8mf2))) +vuint8mf2_t vloxei64(const uint8_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8mf2_m))) +vuint8mf2_t vloxei64(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8mf4))) +vuint8mf4_t vloxei64(const uint8_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8mf4_m))) +vuint8mf4_t vloxei64(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8mf8))) +vuint8mf8_t vloxei64(const uint8_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8mf8_m))) +vuint8mf8_t vloxei64(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m1))) +vint16m1_t vloxei8(const int16_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m1_m))) +vint16m1_t vloxei8(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m2))) +vint16m2_t vloxei8(const int16_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m2_m))) +vint16m2_t vloxei8(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m4))) +vint16m4_t vloxei8(const int16_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m4_m))) +vint16m4_t vloxei8(vbool4_t op0, vint16m4_t op1, const int16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m8))) +vint16m8_t vloxei8(const int16_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m8_m))) +vint16m8_t vloxei8(vbool2_t op0, vint16m8_t op1, const int16_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16mf2))) +vint16mf2_t vloxei8(const int16_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16mf2_m))) +vint16mf2_t vloxei8(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16mf4))) +vint16mf4_t vloxei8(const int16_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16mf4_m))) +vint16mf4_t vloxei8(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m1))) +vuint16m1_t vloxei8(const uint16_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m1_m))) +vuint16m1_t vloxei8(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m2))) +vuint16m2_t vloxei8(const uint16_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m2_m))) +vuint16m2_t vloxei8(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m4))) +vuint16m4_t vloxei8(const uint16_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m4_m))) +vuint16m4_t vloxei8(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m8))) +vuint16m8_t vloxei8(const uint16_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m8_m))) +vuint16m8_t vloxei8(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16mf2))) +vuint16mf2_t vloxei8(const uint16_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16mf2_m))) +vuint16mf2_t vloxei8(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16mf4))) +vuint16mf4_t vloxei8(const uint16_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16mf4_m))) +vuint16mf4_t vloxei8(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m1))) +vint16m1_t vloxei16(const int16_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m1_m))) +vint16m1_t vloxei16(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m2))) +vint16m2_t vloxei16(const int16_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m2_m))) +vint16m2_t vloxei16(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m4))) +vint16m4_t vloxei16(const int16_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m4_m))) +vint16m4_t vloxei16(vbool4_t op0, vint16m4_t op1, const int16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m8))) +vint16m8_t vloxei16(const int16_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m8_m))) +vint16m8_t vloxei16(vbool2_t op0, vint16m8_t op1, const int16_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16mf2))) +vint16mf2_t vloxei16(const int16_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16mf2_m))) +vint16mf2_t vloxei16(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16mf4))) +vint16mf4_t vloxei16(const int16_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16mf4_m))) +vint16mf4_t vloxei16(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m1))) +vuint16m1_t vloxei16(const uint16_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m1_m))) +vuint16m1_t vloxei16(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m2))) +vuint16m2_t vloxei16(const uint16_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m2_m))) +vuint16m2_t vloxei16(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m4))) +vuint16m4_t vloxei16(const uint16_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m4_m))) +vuint16m4_t vloxei16(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m8))) +vuint16m8_t vloxei16(const uint16_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m8_m))) +vuint16m8_t vloxei16(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16mf2))) +vuint16mf2_t vloxei16(const uint16_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16mf2_m))) +vuint16mf2_t vloxei16(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16mf4))) +vuint16mf4_t vloxei16(const uint16_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16mf4_m))) +vuint16mf4_t vloxei16(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16m1))) +vint16m1_t vloxei32(const int16_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16m1_m))) +vint16m1_t vloxei32(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16m2))) +vint16m2_t vloxei32(const int16_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16m2_m))) +vint16m2_t vloxei32(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16m4))) +vint16m4_t vloxei32(const int16_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16m4_m))) +vint16m4_t vloxei32(vbool4_t op0, vint16m4_t op1, const int16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16mf2))) +vint16mf2_t vloxei32(const int16_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16mf2_m))) +vint16mf2_t vloxei32(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16mf4))) +vint16mf4_t vloxei32(const int16_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16mf4_m))) +vint16mf4_t vloxei32(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16m1))) +vuint16m1_t vloxei32(const uint16_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16m1_m))) +vuint16m1_t vloxei32(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16m2))) +vuint16m2_t vloxei32(const uint16_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16m2_m))) +vuint16m2_t vloxei32(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16m4))) +vuint16m4_t vloxei32(const uint16_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16m4_m))) +vuint16m4_t vloxei32(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16mf2))) +vuint16mf2_t vloxei32(const uint16_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16mf2_m))) +vuint16mf2_t vloxei32(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16mf4))) +vuint16mf4_t vloxei32(const uint16_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16mf4_m))) +vuint16mf4_t vloxei32(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16m1))) +vint16m1_t vloxei64(const int16_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16m1_m))) +vint16m1_t vloxei64(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16m2))) +vint16m2_t vloxei64(const int16_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16m2_m))) +vint16m2_t vloxei64(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16mf2))) +vint16mf2_t vloxei64(const int16_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16mf2_m))) +vint16mf2_t vloxei64(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16mf4))) +vint16mf4_t vloxei64(const int16_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16mf4_m))) +vint16mf4_t vloxei64(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16m1))) +vuint16m1_t vloxei64(const uint16_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16m1_m))) +vuint16m1_t vloxei64(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16m2))) +vuint16m2_t vloxei64(const uint16_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16m2_m))) +vuint16m2_t vloxei64(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16mf2))) +vuint16mf2_t vloxei64(const uint16_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16mf2_m))) +vuint16mf2_t vloxei64(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16mf4))) +vuint16mf4_t vloxei64(const uint16_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16mf4_m))) +vuint16mf4_t vloxei64(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m1))) +vint32m1_t vloxei8(const int32_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m1_m))) +vint32m1_t vloxei8(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m2))) +vint32m2_t vloxei8(const int32_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m2_m))) +vint32m2_t vloxei8(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m4))) +vint32m4_t vloxei8(const int32_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m4_m))) +vint32m4_t vloxei8(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m8))) +vint32m8_t vloxei8(const int32_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m8_m))) +vint32m8_t vloxei8(vbool4_t op0, vint32m8_t op1, const int32_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32mf2))) +vint32mf2_t vloxei8(const int32_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32mf2_m))) +vint32mf2_t vloxei8(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m1))) +vuint32m1_t vloxei8(const uint32_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m1_m))) +vuint32m1_t vloxei8(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m2))) +vuint32m2_t vloxei8(const uint32_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m2_m))) +vuint32m2_t vloxei8(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m4))) +vuint32m4_t vloxei8(const uint32_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m4_m))) +vuint32m4_t vloxei8(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m8))) +vuint32m8_t vloxei8(const uint32_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m8_m))) +vuint32m8_t vloxei8(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32mf2))) +vuint32mf2_t vloxei8(const uint32_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32mf2_m))) +vuint32mf2_t vloxei8(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m1))) +vint32m1_t vloxei16(const int32_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m1_m))) +vint32m1_t vloxei16(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m2))) +vint32m2_t vloxei16(const int32_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m2_m))) +vint32m2_t vloxei16(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m4))) +vint32m4_t vloxei16(const int32_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m4_m))) +vint32m4_t vloxei16(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m8))) +vint32m8_t vloxei16(const int32_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m8_m))) +vint32m8_t vloxei16(vbool4_t op0, vint32m8_t op1, const int32_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32mf2))) +vint32mf2_t vloxei16(const int32_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32mf2_m))) +vint32mf2_t vloxei16(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m1))) +vuint32m1_t vloxei16(const uint32_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m1_m))) +vuint32m1_t vloxei16(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m2))) +vuint32m2_t vloxei16(const uint32_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m2_m))) +vuint32m2_t vloxei16(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m4))) +vuint32m4_t vloxei16(const uint32_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m4_m))) +vuint32m4_t vloxei16(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m8))) +vuint32m8_t vloxei16(const uint32_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m8_m))) +vuint32m8_t vloxei16(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32mf2))) +vuint32mf2_t vloxei16(const uint32_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32mf2_m))) +vuint32mf2_t vloxei16(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m1))) +vint32m1_t vloxei32(const int32_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m1_m))) +vint32m1_t vloxei32(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m2))) +vint32m2_t vloxei32(const int32_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m2_m))) +vint32m2_t vloxei32(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m4))) +vint32m4_t vloxei32(const int32_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m4_m))) +vint32m4_t vloxei32(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m8))) +vint32m8_t vloxei32(const int32_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m8_m))) +vint32m8_t vloxei32(vbool4_t op0, vint32m8_t op1, const int32_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32mf2))) +vint32mf2_t vloxei32(const int32_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32mf2_m))) +vint32mf2_t vloxei32(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m1))) +vuint32m1_t vloxei32(const uint32_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m1_m))) +vuint32m1_t vloxei32(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m2))) +vuint32m2_t vloxei32(const uint32_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m2_m))) +vuint32m2_t vloxei32(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m4))) +vuint32m4_t vloxei32(const uint32_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m4_m))) +vuint32m4_t vloxei32(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m8))) +vuint32m8_t vloxei32(const uint32_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m8_m))) +vuint32m8_t vloxei32(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32mf2))) +vuint32mf2_t vloxei32(const uint32_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32mf2_m))) +vuint32mf2_t vloxei32(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32m1))) +vint32m1_t vloxei64(const int32_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32m1_m))) +vint32m1_t vloxei64(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32m2))) +vint32m2_t vloxei64(const int32_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32m2_m))) +vint32m2_t vloxei64(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32m4))) +vint32m4_t vloxei64(const int32_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32m4_m))) +vint32m4_t vloxei64(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32mf2))) +vint32mf2_t vloxei64(const int32_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32mf2_m))) +vint32mf2_t vloxei64(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32m1))) +vuint32m1_t vloxei64(const uint32_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32m1_m))) +vuint32m1_t vloxei64(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32m2))) +vuint32m2_t vloxei64(const uint32_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32m2_m))) +vuint32m2_t vloxei64(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32m4))) +vuint32m4_t vloxei64(const uint32_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32m4_m))) +vuint32m4_t vloxei64(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32mf2))) +vuint32mf2_t vloxei64(const uint32_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32mf2_m))) +vuint32mf2_t vloxei64(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m1))) +vint64m1_t vloxei8(const int64_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m1_m))) +vint64m1_t vloxei8(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m2))) +vint64m2_t vloxei8(const int64_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m2_m))) +vint64m2_t vloxei8(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m4))) +vint64m4_t vloxei8(const int64_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m4_m))) +vint64m4_t vloxei8(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m8))) +vint64m8_t vloxei8(const int64_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m8_m))) +vint64m8_t vloxei8(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m1))) +void vse8(int8_t * op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m1_m))) +void vse8(vbool8_t op0, int8_t * op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m2))) +void vse8(int8_t * op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m2_m))) +void vse8(vbool4_t op0, int8_t * op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m4))) +void vse8(int8_t * op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m4_m))) +void vse8(vbool2_t op0, int8_t * op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m8))) +void vse8(int8_t * op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m8_m))) +void vse8(vbool1_t op0, int8_t * op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8mf2))) +void vse8(int8_t * op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8mf2_m))) +void vse8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8mf4))) +void vse8(int8_t * op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8mf4_m))) +void vse8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8mf8))) +void vse8(int8_t * op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8mf8_m))) +void vse8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m1))) +vuint64m1_t vloxei8(const uint64_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m1_m))) +vuint64m1_t vloxei8(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m2))) +vuint64m2_t vloxei8(const uint64_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m2_m))) +vuint64m2_t vloxei8(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m4))) +vuint64m4_t vloxei8(const uint64_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m4_m))) +vuint64m4_t vloxei8(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m8))) +vuint64m8_t vloxei8(const uint64_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m8_m))) +vuint64m8_t vloxei8(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m1))) +vint64m1_t vloxei16(const int64_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m1_m))) +vint64m1_t vloxei16(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m2))) +vint64m2_t vloxei16(const int64_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m2_m))) +vint64m2_t vloxei16(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m4))) +vint64m4_t vloxei16(const int64_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m4_m))) +vint64m4_t vloxei16(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m8))) +vint64m8_t vloxei16(const int64_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m8_m))) +vint64m8_t vloxei16(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m1))) +vuint64m1_t vloxei16(const uint64_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m1_m))) +vuint64m1_t vloxei16(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m2))) +vuint64m2_t vloxei16(const uint64_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m2_m))) +vuint64m2_t vloxei16(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m4))) +vuint64m4_t vloxei16(const uint64_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m4_m))) +vuint64m4_t vloxei16(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m8))) +vuint64m8_t vloxei16(const uint64_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m8_m))) +vuint64m8_t vloxei16(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m1))) +vint64m1_t vloxei32(const int64_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m1_m))) +vint64m1_t vloxei32(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m2))) +vint64m2_t vloxei32(const int64_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m2_m))) +vint64m2_t vloxei32(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m4))) +vint64m4_t vloxei32(const int64_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m4_m))) +vint64m4_t vloxei32(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m8))) +vint64m8_t vloxei32(const int64_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m8_m))) +vint64m8_t vloxei32(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m1))) +vuint64m1_t vloxei32(const uint64_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m1_m))) +vuint64m1_t vloxei32(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m2))) +vuint64m2_t vloxei32(const uint64_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m2_m))) +vuint64m2_t vloxei32(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m4))) +vuint64m4_t vloxei32(const uint64_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m4_m))) +vuint64m4_t vloxei32(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m8))) +vuint64m8_t vloxei32(const uint64_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m8_m))) +vuint64m8_t vloxei32(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m1))) +vint64m1_t vloxei64(const int64_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m1_m))) +vint64m1_t vloxei64(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m2))) +vint64m2_t vloxei64(const int64_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m2_m))) +vint64m2_t vloxei64(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m4))) +vint64m4_t vloxei64(const int64_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m4_m))) +vint64m4_t vloxei64(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m8))) +vint64m8_t vloxei64(const int64_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m8_m))) +vint64m8_t vloxei64(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m1))) +vuint64m1_t vloxei64(const uint64_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m1_m))) +vuint64m1_t vloxei64(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m2))) +vuint64m2_t vloxei64(const uint64_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m2_m))) +vuint64m2_t vloxei64(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m4))) +vuint64m4_t vloxei64(const uint64_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m4_m))) +vuint64m4_t vloxei64(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m8))) +vuint64m8_t vloxei64(const uint64_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m8_m))) +vuint64m8_t vloxei64(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m1))) +void vse8(uint8_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m1_m))) +void vse8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m2))) +void vse8(uint8_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m2_m))) +void vse8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m4))) +void vse8(uint8_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m4_m))) +void vse8(vbool2_t op0, uint8_t * op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m8))) +void vse8(uint8_t * op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m8_m))) +void vse8(vbool1_t op0, uint8_t * op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8mf2))) +void vse8(uint8_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8mf2_m))) +void vse8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8mf4))) +void vse8(uint8_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8mf4_m))) +void vse8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8mf8))) +void vse8(uint8_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8mf8_m))) +void vse8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16mf4))) +vuint16mf4_t vwaddu_wv(vuint16mf4_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16mf4_m))) +vuint16mf4_t vwaddu_wv(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16mf2))) +vuint16mf2_t vwaddu_wv(vuint16mf2_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16mf2_m))) +vuint16mf2_t vwaddu_wv(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m1))) +vuint16m1_t vwaddu_wv(vuint16m1_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m1_m))) +vuint16m1_t vwaddu_wv(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m2))) +vuint16m2_t vwaddu_wv(vuint16m2_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m2_m))) +vuint16m2_t vwaddu_wv(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m4))) +vuint16m4_t vwaddu_wv(vuint16m4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m4_m))) +vuint16m4_t vwaddu_wv(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m8))) +vuint16m8_t vwaddu_wv(vuint16m8_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m8_m))) +vuint16m8_t vwaddu_wv(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32mf2))) +vuint32mf2_t vwaddu_wv(vuint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32mf2_m))) +vuint32mf2_t vwaddu_wv(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m1))) +vuint32m1_t vwaddu_wv(vuint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m1_m))) +vuint32m1_t vwaddu_wv(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m2))) +vuint32m2_t vwaddu_wv(vuint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m2_m))) +vuint32m2_t vwaddu_wv(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m4))) +vuint32m4_t vwaddu_wv(vuint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m4_m))) +vuint32m4_t vwaddu_wv(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m8))) +vuint32m8_t vwaddu_wv(vuint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m8_m))) +vuint32m8_t vwaddu_wv(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m1))) +vuint64m1_t vwaddu_wv(vuint64m1_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m1_m))) +vuint64m1_t vwaddu_wv(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m2))) +vuint64m2_t vwaddu_wv(vuint64m2_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m2_m))) +vuint64m2_t vwaddu_wv(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m4))) +vuint64m4_t vwaddu_wv(vuint64m4_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m4_m))) +vuint64m4_t vwaddu_wv(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m8))) +vuint64m8_t vwaddu_wv(vuint64m8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m8_m))) +vuint64m8_t vwaddu_wv(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m1))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m1_m))) +void vsse8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m2))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m2_m))) +void vsse8(vbool4_t op0, int8_t * op1, ptrdiff_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m4))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m4_m))) +void vsse8(vbool2_t op0, int8_t * op1, ptrdiff_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m8))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m8_m))) +void vsse8(vbool1_t op0, int8_t * op1, ptrdiff_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8mf2))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8mf2_m))) +void vsse8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8mf4))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8mf4_m))) +void vsse8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8mf8))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8mf8_m))) +void vsse8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8m1))) +void vsuxei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8m1_m))) +void vsuxei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8m2))) +void vsuxei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8m2_m))) +void vsuxei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8m4))) +void vsuxei16(int8_t * op0, vuint16m8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8m4_m))) +void vsuxei16(vbool2_t op0, int8_t * op1, vuint16m8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8mf2))) +void vsuxei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8mf2_m))) +void vsuxei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8mf4))) +void vsuxei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8mf4_m))) +void vsuxei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8mf8))) +void vsuxei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8mf8_m))) +void vsuxei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8m1))) +void vsuxei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8m1_m))) +void vsuxei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8m2))) +void vsuxei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8m2_m))) +void vsuxei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8m4))) +void vsuxei16(uint8_t * op0, vuint16m8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8m4_m))) +void vsuxei16(vbool2_t op0, uint8_t * op1, vuint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8mf2))) +void vsuxei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8mf2_m))) +void vsuxei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8mf4))) +void vsuxei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8mf4_m))) +void vsuxei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8mf8))) +void vsuxei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8mf8_m))) +void vsuxei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8m1))) +void vsuxei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8m1_m))) +void vsuxei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8m2))) +void vsuxei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8m2_m))) +void vsuxei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8mf2))) +void vsuxei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8mf2_m))) +void vsuxei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8mf4))) +void vsuxei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8mf4_m))) +void vsuxei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8mf8))) +void vsuxei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8mf8_m))) +void vsuxei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m1))) +vint8m1_t vadd(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m1_m))) +vint8m1_t vadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m2))) +vint8m2_t vadd(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m2_m))) +vint8m2_t vadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m4))) +vint8m4_t vadd(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m4_m))) +vint8m4_t vadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m8))) +vint8m8_t vadd(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m8_m))) +vint8m8_t vadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8mf2))) +vint8mf2_t vadd(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8mf2_m))) +vint8mf2_t vadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8mf4))) +vint8mf4_t vadd(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8mf4_m))) +vint8mf4_t vadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8mf8))) +vint8mf8_t vadd(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8mf8_m))) +vint8mf8_t vadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m1))) +vint16m1_t vadd(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m1_m))) +vint16m1_t vadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m2))) +vint16m2_t vadd(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m2_m))) +vint16m2_t vadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m4))) +vint16m4_t vadd(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m4_m))) +vint16m4_t vadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m8))) +vint16m8_t vadd(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m8_m))) +vint16m8_t vadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16mf2))) +vint16mf2_t vadd(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16mf2_m))) +vint16mf2_t vadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16mf4))) +vint16mf4_t vadd(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16mf4_m))) +vint16mf4_t vadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m1))) +vint32m1_t vadd(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m1_m))) +vint32m1_t vadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m2))) +vint32m2_t vadd(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m2_m))) +vint32m2_t vadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m4))) +vint32m4_t vadd(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m4_m))) +vint32m4_t vadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m8))) +vint32m8_t vadd(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m8_m))) +vint32m8_t vadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32mf2))) +vint32mf2_t vadd(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32mf2_m))) +vint32mf2_t vadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m1))) +vint64m1_t vadd(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m1_m))) +vint64m1_t vadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m2))) +vint64m2_t vadd(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m2_m))) +vint64m2_t vadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m4))) +vint64m4_t vadd(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m4_m))) +vint64m4_t vadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m8))) +vint64m8_t vadd(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m8_m))) +vint64m8_t vadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m1))) +vuint8m1_t vadd(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m1_m))) +vuint8m1_t vadd(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m2))) +vuint8m2_t vadd(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m2_m))) +vuint8m2_t vadd(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m4))) +vuint8m4_t vadd(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m4_m))) +vuint8m4_t vadd(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m8))) +vuint8m8_t vadd(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m8_m))) +vuint8m8_t vadd(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8mf2))) +vuint8mf2_t vadd(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8mf2_m))) +vuint8mf2_t vadd(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8mf4))) +vuint8mf4_t vadd(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8mf4_m))) +vuint8mf4_t vadd(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8mf8))) +vuint8mf8_t vadd(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8mf8_m))) +vuint8mf8_t vadd(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m1))) +vuint16m1_t vadd(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m1_m))) +vuint16m1_t vadd(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m2))) +vuint16m2_t vadd(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m2_m))) +vuint16m2_t vadd(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m4))) +vuint16m4_t vadd(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m4_m))) +vuint16m4_t vadd(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m8))) +vuint16m8_t vadd(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m8_m))) +vuint16m8_t vadd(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16mf2))) +vuint16mf2_t vadd(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16mf2_m))) +vuint16mf2_t vadd(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16mf4))) +vuint16mf4_t vadd(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16mf4_m))) +vuint16mf4_t vadd(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m1))) +vuint32m1_t vadd(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m1_m))) +vuint32m1_t vadd(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m2))) +vuint32m2_t vadd(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m2_m))) +vuint32m2_t vadd(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m4))) +vuint32m4_t vadd(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m4_m))) +vuint32m4_t vadd(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m8))) +vuint32m8_t vadd(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m8_m))) +vuint32m8_t vadd(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32mf2))) +vuint32mf2_t vadd(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32mf2_m))) +vuint32mf2_t vadd(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m1))) +vuint64m1_t vadd(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m1_m))) +vuint64m1_t vadd(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m2))) +vuint64m2_t vadd(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m2_m))) +vuint64m2_t vadd(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m4))) +vuint64m4_t vadd(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m4_m))) +vuint64m4_t vadd(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m8))) +vuint64m8_t vadd(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m8_m))) +vuint64m8_t vadd(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m1))) +vuint8m1_t vadd(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m1_m))) +vuint8m1_t vadd(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m2))) +vuint8m2_t vadd(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m2_m))) +vuint8m2_t vadd(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m4))) +vuint8m4_t vadd(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m4_m))) +vuint8m4_t vadd(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m8))) +vuint8m8_t vadd(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m8_m))) +vuint8m8_t vadd(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8mf2))) +vuint8mf2_t vadd(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8mf2_m))) +vuint8mf2_t vadd(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8mf4))) +vuint8mf4_t vadd(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8mf4_m))) +vuint8mf4_t vadd(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8mf8))) +vuint8mf8_t vadd(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8mf8_m))) +vuint8mf8_t vadd(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m1))) +vuint16m1_t vadd(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m1_m))) +vuint16m1_t vadd(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m2))) +vuint16m2_t vadd(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m2_m))) +vuint16m2_t vadd(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m4))) +vuint16m4_t vadd(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m4_m))) +vuint16m4_t vadd(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m8))) +vuint16m8_t vadd(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m8_m))) +vuint16m8_t vadd(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16mf2))) +vuint16mf2_t vadd(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16mf2_m))) +vuint16mf2_t vadd(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16mf4))) +vuint16mf4_t vadd(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16mf4_m))) +vuint16mf4_t vadd(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m1))) +vuint32m1_t vadd(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m1_m))) +vuint32m1_t vadd(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m2))) +vuint32m2_t vadd(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m2_m))) +vuint32m2_t vadd(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m4))) +vuint32m4_t vadd(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m4_m))) +vuint32m4_t vadd(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m8))) +vuint32m8_t vadd(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m8_m))) +vuint32m8_t vadd(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32mf2))) +vuint32mf2_t vadd(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32mf2_m))) +vuint32mf2_t vadd(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m1))) +vuint64m1_t vadd(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m1_m))) +vuint64m1_t vadd(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m2))) +vuint64m2_t vadd(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m2_m))) +vuint64m2_t vadd(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m4))) +vuint64m4_t vadd(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m4_m))) +vuint64m4_t vadd(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m8))) +vuint64m8_t vadd(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m8_m))) +vuint64m8_t vadd(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m1))) +vint8m1_t vsub(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m1_m))) +vint8m1_t vsub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m2))) +vint8m2_t vsub(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m2_m))) +vint8m2_t vsub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m4))) +vint8m4_t vsub(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m4_m))) +vint8m4_t vsub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m8))) +vint8m8_t vsub(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m8_m))) +vint8m8_t vsub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8mf2))) +vint8mf2_t vsub(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8mf2_m))) +vint8mf2_t vsub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8mf4))) +vint8mf4_t vsub(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8mf4_m))) +vint8mf4_t vsub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8mf8))) +vint8mf8_t vsub(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8mf8_m))) +vint8mf8_t vsub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m1))) +vint16m1_t vsub(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m1_m))) +vint16m1_t vsub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m2))) +vint16m2_t vsub(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m2_m))) +vint16m2_t vsub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m4))) +vint16m4_t vsub(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m4_m))) +vint16m4_t vsub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m8))) +vint16m8_t vsub(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m8_m))) +vint16m8_t vsub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16mf2))) +vint16mf2_t vsub(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16mf2_m))) +vint16mf2_t vsub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16mf4))) +vint16mf4_t vsub(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16mf4_m))) +vint16mf4_t vsub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m1))) +vint32m1_t vsub(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m1_m))) +vint32m1_t vsub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m2))) +vint32m2_t vsub(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m2_m))) +vint32m2_t vsub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m4))) +vint32m4_t vsub(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m4_m))) +vint32m4_t vsub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m8))) +vint32m8_t vsub(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m8_m))) +vint32m8_t vsub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32mf2))) +vint32mf2_t vsub(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32mf2_m))) +vint32mf2_t vsub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m1))) +vint64m1_t vsub(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m1_m))) +vint64m1_t vsub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m2))) +vint64m2_t vsub(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m2_m))) +vint64m2_t vsub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m4))) +vint64m4_t vsub(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m4_m))) +vint64m4_t vsub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m8))) +vint64m8_t vsub(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m8_m))) +vint64m8_t vsub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8m1))) +void vsuxei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8m1_m))) +void vsuxei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8m2))) +void vsuxei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8m2_m))) +void vsuxei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8mf2))) +void vsuxei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8mf2_m))) +void vsuxei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8mf4))) +void vsuxei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8mf4_m))) +void vsuxei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8mf8))) +void vsuxei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8mf8_m))) +void vsuxei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m1))) +vint8m1_t vsub(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m1_m))) +vint8m1_t vsub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m2))) +vint8m2_t vsub(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m2_m))) +vint8m2_t vsub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m4))) +vint8m4_t vsub(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m4_m))) +vint8m4_t vsub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m8))) +vint8m8_t vsub(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m8_m))) +vint8m8_t vsub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8mf2))) +vint8mf2_t vsub(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8mf2_m))) +vint8mf2_t vsub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8mf4))) +vint8mf4_t vsub(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8mf4_m))) +vint8mf4_t vsub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8mf8))) +vint8mf8_t vsub(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8mf8_m))) +vint8mf8_t vsub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m1))) +vint16m1_t vsub(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m1_m))) +vint16m1_t vsub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m2))) +vint16m2_t vsub(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m2_m))) +vint16m2_t vsub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m4))) +vint16m4_t vsub(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m4_m))) +vint16m4_t vsub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m8))) +vint16m8_t vsub(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m8_m))) +vint16m8_t vsub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16mf2))) +vint16mf2_t vsub(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16mf2_m))) +vint16mf2_t vsub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16mf4))) +vint16mf4_t vsub(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16mf4_m))) +vint16mf4_t vsub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m1))) +vint32m1_t vsub(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m1_m))) +vint32m1_t vsub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m2))) +vint32m2_t vsub(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m2_m))) +vint32m2_t vsub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m4))) +vint32m4_t vsub(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m4_m))) +vint32m4_t vsub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m8))) +vint32m8_t vsub(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m8_m))) +vint32m8_t vsub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32mf2))) +vint32mf2_t vsub(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32mf2_m))) +vint32mf2_t vsub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m1))) +vint64m1_t vsub(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m1_m))) +vint64m1_t vsub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m2))) +vint64m2_t vsub(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m2_m))) +vint64m2_t vsub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m4))) +vint64m4_t vsub(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m4_m))) +vint64m4_t vsub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m8))) +vint64m8_t vsub(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m8_m))) +vint64m8_t vsub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m1))) +vuint8m1_t vsub(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m1_m))) +vuint8m1_t vsub(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m2))) +vuint8m2_t vsub(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m2_m))) +vuint8m2_t vsub(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m4))) +vuint8m4_t vsub(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m4_m))) +vuint8m4_t vsub(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m8))) +vuint8m8_t vsub(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m8_m))) +vuint8m8_t vsub(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8mf2))) +vuint8mf2_t vsub(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8mf2_m))) +vuint8mf2_t vsub(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8mf4))) +vuint8mf4_t vsub(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8mf4_m))) +vuint8mf4_t vsub(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8mf8))) +vuint8mf8_t vsub(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8mf8_m))) +vuint8mf8_t vsub(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m1))) +vuint16m1_t vsub(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m1_m))) +vuint16m1_t vsub(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m2))) +vuint16m2_t vsub(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m2_m))) +vuint16m2_t vsub(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m4))) +vuint16m4_t vsub(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m4_m))) +vuint16m4_t vsub(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m8))) +vuint16m8_t vsub(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m8_m))) +vuint16m8_t vsub(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16mf2))) +vuint16mf2_t vsub(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16mf2_m))) +vuint16mf2_t vsub(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16mf4))) +vuint16mf4_t vsub(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16mf4_m))) +vuint16mf4_t vsub(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m1))) +vuint32m1_t vsub(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m1_m))) +vuint32m1_t vsub(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m2))) +vuint32m2_t vsub(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m2_m))) +vuint32m2_t vsub(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m4))) +vuint32m4_t vsub(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m4_m))) +vuint32m4_t vsub(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m8))) +vuint32m8_t vsub(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m8_m))) +vuint32m8_t vsub(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32mf2))) +vuint32mf2_t vsub(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32mf2_m))) +vuint32mf2_t vsub(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m1))) +vuint64m1_t vsub(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m1_m))) +vuint64m1_t vsub(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m2))) +vuint64m2_t vsub(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m2_m))) +vuint64m2_t vsub(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m4))) +vuint64m4_t vsub(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m4_m))) +vuint64m4_t vsub(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m8))) +vuint64m8_t vsub(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m8_m))) +vuint64m8_t vsub(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m1))) +vuint8m1_t vsub(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m1_m))) +vuint8m1_t vsub(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m2))) +vuint8m2_t vsub(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m2_m))) +vuint8m2_t vsub(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m4))) +vuint8m4_t vsub(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m4_m))) +vuint8m4_t vsub(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m8))) +vuint8m8_t vsub(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m8_m))) +vuint8m8_t vsub(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8mf2))) +vuint8mf2_t vsub(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8mf2_m))) +vuint8mf2_t vsub(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8mf4))) +vuint8mf4_t vsub(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8mf4_m))) +vuint8mf4_t vsub(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8mf8))) +vuint8mf8_t vsub(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8mf8_m))) +vuint8mf8_t vsub(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m1))) +vuint16m1_t vsub(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m1_m))) +vuint16m1_t vsub(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m2))) +vuint16m2_t vsub(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m2_m))) +vuint16m2_t vsub(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m4))) +vuint16m4_t vsub(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m4_m))) +vuint16m4_t vsub(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m8))) +vuint16m8_t vsub(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m8_m))) +vuint16m8_t vsub(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16mf2))) +vuint16mf2_t vsub(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16mf2_m))) +vuint16mf2_t vsub(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16mf4))) +vuint16mf4_t vsub(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16mf4_m))) +vuint16mf4_t vsub(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m1))) +vuint32m1_t vsub(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m1_m))) +vuint32m1_t vsub(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m2))) +vuint32m2_t vsub(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m2_m))) +vuint32m2_t vsub(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m4))) +vuint32m4_t vsub(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m4_m))) +vuint32m4_t vsub(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m8))) +vuint32m8_t vsub(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m8_m))) +vuint32m8_t vsub(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32mf2))) +vuint32mf2_t vsub(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32mf2_m))) +vuint32mf2_t vsub(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m1))) +vuint64m1_t vsub(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m1_m))) +vuint64m1_t vsub(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m2))) +vuint64m2_t vsub(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m2_m))) +vuint64m2_t vsub(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m4))) +vuint64m4_t vsub(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m4_m))) +vuint64m4_t vsub(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m8))) +vuint64m8_t vsub(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m8_m))) +vuint64m8_t vsub(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m1))) +vint8m1_t vrsub(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m1_m))) +vint8m1_t vrsub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m2))) +vint8m2_t vrsub(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m2_m))) +vint8m2_t vrsub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m4))) +vint8m4_t vrsub(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m4_m))) +vint8m4_t vrsub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m8))) +vint8m8_t vrsub(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m8_m))) +vint8m8_t vrsub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8mf2))) +vint8mf2_t vrsub(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8mf2_m))) +vint8mf2_t vrsub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8mf4))) +vint8mf4_t vrsub(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8mf4_m))) +vint8mf4_t vrsub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8mf8))) +vint8mf8_t vrsub(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8mf8_m))) +vint8mf8_t vrsub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m1))) +vint16m1_t vrsub(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m1_m))) +vint16m1_t vrsub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m2))) +vint16m2_t vrsub(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m2_m))) +vint16m2_t vrsub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m4))) +vint16m4_t vrsub(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m4_m))) +vint16m4_t vrsub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m8))) +vint16m8_t vrsub(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m8_m))) +vint16m8_t vrsub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16mf2))) +vint16mf2_t vrsub(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16mf2_m))) +vint16mf2_t vrsub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16mf4))) +vint16mf4_t vrsub(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16mf4_m))) +vint16mf4_t vrsub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m1))) +vint32m1_t vrsub(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m1_m))) +vint32m1_t vrsub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m2))) +vint32m2_t vrsub(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m2_m))) +vint32m2_t vrsub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m4))) +vint32m4_t vrsub(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m4_m))) +vint32m4_t vrsub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m8))) +vint32m8_t vrsub(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m8_m))) +vint32m8_t vrsub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32mf2))) +vint32mf2_t vrsub(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32mf2_m))) +vint32mf2_t vrsub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m1))) +vint64m1_t vrsub(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m1_m))) +vint64m1_t vrsub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m2))) +vint64m2_t vrsub(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m2_m))) +vint64m2_t vrsub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m4))) +vint64m4_t vrsub(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m4_m))) +vint64m4_t vrsub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m8))) +vint64m8_t vrsub(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m8_m))) +vint64m8_t vrsub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m1))) +vuint8m1_t vrsub(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m1_m))) +vuint8m1_t vrsub(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m2))) +vuint8m2_t vrsub(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m2_m))) +vuint8m2_t vrsub(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m4))) +vuint8m4_t vrsub(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m4_m))) +vuint8m4_t vrsub(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m8))) +vuint8m8_t vrsub(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m8_m))) +vuint8m8_t vrsub(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8mf2))) +vuint8mf2_t vrsub(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8mf2_m))) +vuint8mf2_t vrsub(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8mf4))) +vuint8mf4_t vrsub(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8mf4_m))) +vuint8mf4_t vrsub(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8mf8))) +vuint8mf8_t vrsub(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8mf8_m))) +vuint8mf8_t vrsub(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m1))) +vuint16m1_t vrsub(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m1_m))) +vuint16m1_t vrsub(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m2))) +vuint16m2_t vrsub(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m2_m))) +vuint16m2_t vrsub(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m4))) +vuint16m4_t vrsub(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m4_m))) +vuint16m4_t vrsub(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m8))) +vuint16m8_t vrsub(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m8_m))) +vuint16m8_t vrsub(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16mf2))) +vuint16mf2_t vrsub(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16mf2_m))) +vuint16mf2_t vrsub(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16mf4))) +vuint16mf4_t vrsub(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16mf4_m))) +vuint16mf4_t vrsub(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m1))) +vuint32m1_t vrsub(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m1_m))) +vuint32m1_t vrsub(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m2))) +vuint32m2_t vrsub(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m2_m))) +vuint32m2_t vrsub(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m4))) +vuint32m4_t vrsub(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m4_m))) +vuint32m4_t vrsub(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m8))) +vuint32m8_t vrsub(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m8_m))) +vuint32m8_t vrsub(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32mf2))) +vuint32mf2_t vrsub(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32mf2_m))) +vuint32mf2_t vrsub(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m1))) +vuint64m1_t vrsub(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m1_m))) +vuint64m1_t vrsub(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m2))) +vuint64m2_t vrsub(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m2_m))) +vuint64m2_t vrsub(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m4))) +vuint64m4_t vrsub(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m4_m))) +vuint64m4_t vrsub(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m8))) +vuint64m8_t vrsub(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m8_m))) +vuint64m8_t vrsub(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16mf4))) +vuint16mf4_t vwaddu_vx(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16mf4_m))) +vuint16mf4_t vwaddu_vx(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16mf2))) +vuint16mf2_t vwaddu_vx(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16mf2_m))) +vuint16mf2_t vwaddu_vx(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m1))) +vuint16m1_t vwaddu_vx(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m1_m))) +vuint16m1_t vwaddu_vx(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m2))) +vuint16m2_t vwaddu_vx(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m2_m))) +vuint16m2_t vwaddu_vx(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m4))) +vuint16m4_t vwaddu_vx(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m4_m))) +vuint16m4_t vwaddu_vx(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m8))) +vuint16m8_t vwaddu_vx(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m8_m))) +vuint16m8_t vwaddu_vx(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32mf2))) +vuint32mf2_t vwaddu_vx(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32mf2_m))) +vuint32mf2_t vwaddu_vx(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m1))) +vuint32m1_t vwaddu_vx(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m1_m))) +vuint32m1_t vwaddu_vx(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m2))) +vuint32m2_t vwaddu_vx(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m2_m))) +vuint32m2_t vwaddu_vx(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m4))) +vuint32m4_t vwaddu_vx(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m4_m))) +vuint32m4_t vwaddu_vx(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m8))) +vuint32m8_t vwaddu_vx(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m8_m))) +vuint32m8_t vwaddu_vx(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m1))) +vuint64m1_t vwaddu_vx(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m1_m))) +vuint64m1_t vwaddu_vx(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m2))) +vuint64m2_t vwaddu_vx(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m2_m))) +vuint64m2_t vwaddu_vx(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m4))) +vuint64m4_t vwaddu_vx(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m4_m))) +vuint64m4_t vwaddu_vx(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m8))) +vuint64m8_t vwaddu_vx(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m8_m))) +vuint64m8_t vwaddu_vx(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16mf4))) +vuint16mf4_t vwsubu_vv(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16mf4_m))) +vuint16mf4_t vwsubu_vv(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16mf2))) +vuint16mf2_t vwsubu_vv(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16mf2_m))) +vuint16mf2_t vwsubu_vv(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m1))) +vuint16m1_t vwsubu_vv(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m1_m))) +vuint16m1_t vwsubu_vv(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m2))) +vuint16m2_t vwsubu_vv(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m2_m))) +vuint16m2_t vwsubu_vv(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m4))) +vuint16m4_t vwsubu_vv(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m4_m))) +vuint16m4_t vwsubu_vv(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m8))) +vuint16m8_t vwsubu_vv(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m8_m))) +vuint16m8_t vwsubu_vv(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32mf2))) +vuint32mf2_t vwsubu_vv(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32mf2_m))) +vuint32mf2_t vwsubu_vv(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m1))) +vuint32m1_t vwsubu_vv(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m1_m))) +vuint32m1_t vwsubu_vv(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m2))) +vuint32m2_t vwsubu_vv(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m2_m))) +vuint32m2_t vwsubu_vv(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m4))) +vuint32m4_t vwsubu_vv(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m4_m))) +vuint32m4_t vwsubu_vv(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m8))) +vuint32m8_t vwsubu_vv(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m8_m))) +vuint32m8_t vwsubu_vv(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m1))) +vuint64m1_t vwsubu_vv(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m1_m))) +vuint64m1_t vwsubu_vv(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m2))) +vuint64m2_t vwsubu_vv(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m2_m))) +vuint64m2_t vwsubu_vv(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m4))) +vuint64m4_t vwsubu_vv(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m4_m))) +vuint64m4_t vwsubu_vv(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m8))) +vuint64m8_t vwsubu_vv(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m8_m))) +vuint64m8_t vwsubu_vv(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16mf4))) +vuint16mf4_t vwsubu_vx(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16mf4_m))) +vuint16mf4_t vwsubu_vx(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16mf2))) +vuint16mf2_t vwsubu_vx(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16mf2_m))) +vuint16mf2_t vwsubu_vx(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m1))) +vuint16m1_t vwsubu_vx(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m1_m))) +vuint16m1_t vwsubu_vx(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m2))) +vuint16m2_t vwsubu_vx(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m2_m))) +vuint16m2_t vwsubu_vx(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m4))) +vuint16m4_t vwsubu_vx(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m4_m))) +vuint16m4_t vwsubu_vx(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m8))) +vuint16m8_t vwsubu_vx(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m8_m))) +vuint16m8_t vwsubu_vx(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32mf2))) +vuint32mf2_t vwsubu_vx(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32mf2_m))) +vuint32mf2_t vwsubu_vx(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m1))) +vuint32m1_t vwsubu_vx(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m1_m))) +vuint32m1_t vwsubu_vx(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m2))) +vuint32m2_t vwsubu_vx(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m2_m))) +vuint32m2_t vwsubu_vx(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m4))) +vuint32m4_t vwsubu_vx(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m4_m))) +vuint32m4_t vwsubu_vx(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m8))) +vuint32m8_t vwsubu_vx(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m8_m))) +vuint32m8_t vwsubu_vx(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m1))) +vuint64m1_t vwsubu_vx(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m1_m))) +vuint64m1_t vwsubu_vx(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m2))) +vuint64m2_t vwsubu_vx(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m2_m))) +vuint64m2_t vwsubu_vx(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m4))) +vuint64m4_t vwsubu_vx(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m4_m))) +vuint64m4_t vwsubu_vx(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m8))) +vuint64m8_t vwsubu_vx(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m8_m))) +vuint64m8_t vwsubu_vx(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16mf4))) +vint16mf4_t vwadd_vv(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16mf4_m))) +vint16mf4_t vwadd_vv(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16mf2))) +vint16mf2_t vwadd_vv(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16mf2_m))) +vint16mf2_t vwadd_vv(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m1))) +vint16m1_t vwadd_vv(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m1_m))) +vint16m1_t vwadd_vv(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m2))) +vint16m2_t vwadd_vv(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m2_m))) +vint16m2_t vwadd_vv(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m4))) +vint16m4_t vwadd_vv(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m4_m))) +vint16m4_t vwadd_vv(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m8))) +vint16m8_t vwadd_vv(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m8_m))) +vint16m8_t vwadd_vv(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32mf2))) +vint32mf2_t vwadd_vv(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32mf2_m))) +vint32mf2_t vwadd_vv(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m1))) +vint32m1_t vwadd_vv(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m1_m))) +vint32m1_t vwadd_vv(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m2))) +vint32m2_t vwadd_vv(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m2_m))) +vint32m2_t vwadd_vv(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m4))) +vint32m4_t vwadd_vv(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m4_m))) +vint32m4_t vwadd_vv(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m8))) +vint32m8_t vwadd_vv(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m8_m))) +vint32m8_t vwadd_vv(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m1))) +vint64m1_t vwadd_vv(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m1_m))) +vint64m1_t vwadd_vv(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m2))) +vint64m2_t vwadd_vv(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m2_m))) +vint64m2_t vwadd_vv(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m4))) +vint64m4_t vwadd_vv(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m4_m))) +vint64m4_t vwadd_vv(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m8))) +vint64m8_t vwadd_vv(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m8_m))) +vint64m8_t vwadd_vv(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16mf4))) +vint16mf4_t vwadd_vx(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16mf4_m))) +vint16mf4_t vwadd_vx(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16mf2))) +vint16mf2_t vwadd_vx(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16mf2_m))) +vint16mf2_t vwadd_vx(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m1))) +vint16m1_t vwadd_vx(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m1_m))) +vint16m1_t vwadd_vx(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m2))) +vint16m2_t vwadd_vx(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m2_m))) +vint16m2_t vwadd_vx(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m4))) +vint16m4_t vwadd_vx(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m4_m))) +vint16m4_t vwadd_vx(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m8))) +vint16m8_t vwadd_vx(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m8_m))) +vint16m8_t vwadd_vx(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32mf2))) +vint32mf2_t vwadd_vx(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32mf2_m))) +vint32mf2_t vwadd_vx(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m1))) +vint32m1_t vwadd_vx(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m1_m))) +vint32m1_t vwadd_vx(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m2))) +vint32m2_t vwadd_vx(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m2_m))) +vint32m2_t vwadd_vx(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m4))) +vint32m4_t vwadd_vx(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m4_m))) +vint32m4_t vwadd_vx(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m8))) +vint32m8_t vwadd_vx(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m8_m))) +vint32m8_t vwadd_vx(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m1))) +vint64m1_t vwadd_vx(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m1_m))) +vint64m1_t vwadd_vx(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m2))) +vint64m2_t vwadd_vx(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m2_m))) +vint64m2_t vwadd_vx(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m4))) +vint64m4_t vwadd_vx(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m4_m))) +vint64m4_t vwadd_vx(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m8))) +vint64m8_t vwadd_vx(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m8_m))) +vint64m8_t vwadd_vx(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8m1))) +void vsuxei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8m1_m))) +void vsuxei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8mf2))) +void vsuxei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8mf2_m))) +void vsuxei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8mf4))) +void vsuxei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8mf4_m))) +void vsuxei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8mf8))) +void vsuxei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8mf8_m))) +void vsuxei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16mf4))) +vint16mf4_t vwsub_vv(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16mf4_m))) +vint16mf4_t vwsub_vv(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16mf2))) +vint16mf2_t vwsub_vv(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16mf2_m))) +vint16mf2_t vwsub_vv(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m1))) +vint16m1_t vwsub_vv(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m1_m))) +vint16m1_t vwsub_vv(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m2))) +vint16m2_t vwsub_vv(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m2_m))) +vint16m2_t vwsub_vv(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m4))) +vint16m4_t vwsub_vv(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m4_m))) +vint16m4_t vwsub_vv(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m8))) +vint16m8_t vwsub_vv(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m8_m))) +vint16m8_t vwsub_vv(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32mf2))) +vint32mf2_t vwsub_vv(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32mf2_m))) +vint32mf2_t vwsub_vv(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m1))) +vint32m1_t vwsub_vv(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m1_m))) +vint32m1_t vwsub_vv(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m2))) +vint32m2_t vwsub_vv(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m2_m))) +vint32m2_t vwsub_vv(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m4))) +vint32m4_t vwsub_vv(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m4_m))) +vint32m4_t vwsub_vv(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m8))) +vint32m8_t vwsub_vv(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m8_m))) +vint32m8_t vwsub_vv(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m1))) +vint64m1_t vwsub_vv(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m1_m))) +vint64m1_t vwsub_vv(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m2))) +vint64m2_t vwsub_vv(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m2_m))) +vint64m2_t vwsub_vv(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m4))) +vint64m4_t vwsub_vv(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m4_m))) +vint64m4_t vwsub_vv(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m8))) +vint64m8_t vwsub_vv(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m8_m))) +vint64m8_t vwsub_vv(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16mf4))) +vint16mf4_t vwsub_vx(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16mf4_m))) +vint16mf4_t vwsub_vx(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16mf2))) +vint16mf2_t vwsub_vx(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16mf2_m))) +vint16mf2_t vwsub_vx(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m1))) +vint16m1_t vwsub_vx(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m1_m))) +vint16m1_t vwsub_vx(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m2))) +vint16m2_t vwsub_vx(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m2_m))) +vint16m2_t vwsub_vx(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m4))) +vint16m4_t vwsub_vx(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m4_m))) +vint16m4_t vwsub_vx(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m8))) +vint16m8_t vwsub_vx(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m8_m))) +vint16m8_t vwsub_vx(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32mf2))) +vint32mf2_t vwsub_vx(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32mf2_m))) +vint32mf2_t vwsub_vx(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m1))) +vint32m1_t vwsub_vx(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m1_m))) +vint32m1_t vwsub_vx(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m2))) +vint32m2_t vwsub_vx(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m2_m))) +vint32m2_t vwsub_vx(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m4))) +vint32m4_t vwsub_vx(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m4_m))) +vint32m4_t vwsub_vx(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m8))) +vint32m8_t vwsub_vx(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m8_m))) +vint32m8_t vwsub_vx(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m1))) +vint64m1_t vwsub_vx(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m1_m))) +vint64m1_t vwsub_vx(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m2))) +vint64m2_t vwsub_vx(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m2_m))) +vint64m2_t vwsub_vx(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m4))) +vint64m4_t vwsub_vx(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m4_m))) +vint64m4_t vwsub_vx(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m8))) +vint64m8_t vwsub_vx(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m8_m))) +vint64m8_t vwsub_vx(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16mf4))) +vuint16mf4_t vwaddu_wx(vuint16mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16mf4_m))) +vuint16mf4_t vwaddu_wx(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16mf2))) +vuint16mf2_t vwaddu_wx(vuint16mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16mf2_m))) +vuint16mf2_t vwaddu_wx(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m1))) +vuint16m1_t vwaddu_wx(vuint16m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m1_m))) +vuint16m1_t vwaddu_wx(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m2))) +vuint16m2_t vwaddu_wx(vuint16m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m2_m))) +vuint16m2_t vwaddu_wx(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m4))) +vuint16m4_t vwaddu_wx(vuint16m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m4_m))) +vuint16m4_t vwaddu_wx(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m8))) +vuint16m8_t vwaddu_wx(vuint16m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m8_m))) +vuint16m8_t vwaddu_wx(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32mf2))) +vuint32mf2_t vwaddu_wx(vuint32mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32mf2_m))) +vuint32mf2_t vwaddu_wx(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m1))) +vuint32m1_t vwaddu_wx(vuint32m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m1_m))) +vuint32m1_t vwaddu_wx(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m2))) +vuint32m2_t vwaddu_wx(vuint32m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m2_m))) +vuint32m2_t vwaddu_wx(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m4))) +vuint32m4_t vwaddu_wx(vuint32m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m4_m))) +vuint32m4_t vwaddu_wx(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m8))) +vuint32m8_t vwaddu_wx(vuint32m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m8_m))) +vuint32m8_t vwaddu_wx(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m1))) +vuint64m1_t vwaddu_wx(vuint64m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m1_m))) +vuint64m1_t vwaddu_wx(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m2))) +vuint64m2_t vwaddu_wx(vuint64m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m2_m))) +vuint64m2_t vwaddu_wx(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m4))) +vuint64m4_t vwaddu_wx(vuint64m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m4_m))) +vuint64m4_t vwaddu_wx(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m8))) +vuint64m8_t vwaddu_wx(vuint64m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m8_m))) +vuint64m8_t vwaddu_wx(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16mf4))) +vuint16mf4_t vwsubu_wv(vuint16mf4_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16mf4_m))) +vuint16mf4_t vwsubu_wv(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16mf2))) +vuint16mf2_t vwsubu_wv(vuint16mf2_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16mf2_m))) +vuint16mf2_t vwsubu_wv(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m1))) +vuint16m1_t vwsubu_wv(vuint16m1_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m1_m))) +vuint16m1_t vwsubu_wv(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m2))) +vuint16m2_t vwsubu_wv(vuint16m2_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m2_m))) +vuint16m2_t vwsubu_wv(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m4))) +vuint16m4_t vwsubu_wv(vuint16m4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m4_m))) +vuint16m4_t vwsubu_wv(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m8))) +vuint16m8_t vwsubu_wv(vuint16m8_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m8_m))) +vuint16m8_t vwsubu_wv(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32mf2))) +vuint32mf2_t vwsubu_wv(vuint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32mf2_m))) +vuint32mf2_t vwsubu_wv(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m1))) +vuint32m1_t vwsubu_wv(vuint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m1_m))) +vuint32m1_t vwsubu_wv(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m2))) +vuint32m2_t vwsubu_wv(vuint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m2_m))) +vuint32m2_t vwsubu_wv(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m4))) +vuint32m4_t vwsubu_wv(vuint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m4_m))) +vuint32m4_t vwsubu_wv(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m8))) +vuint32m8_t vwsubu_wv(vuint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m8_m))) +vuint32m8_t vwsubu_wv(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m1))) +vuint64m1_t vwsubu_wv(vuint64m1_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m1_m))) +vuint64m1_t vwsubu_wv(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m2))) +vuint64m2_t vwsubu_wv(vuint64m2_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m2_m))) +vuint64m2_t vwsubu_wv(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m4))) +vuint64m4_t vwsubu_wv(vuint64m4_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m4_m))) +vuint64m4_t vwsubu_wv(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m8))) +vuint64m8_t vwsubu_wv(vuint64m8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m8_m))) +vuint64m8_t vwsubu_wv(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16mf4))) +vuint16mf4_t vwsubu_wx(vuint16mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16mf4_m))) +vuint16mf4_t vwsubu_wx(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16mf2))) +vuint16mf2_t vwsubu_wx(vuint16mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16mf2_m))) +vuint16mf2_t vwsubu_wx(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m1))) +vuint16m1_t vwsubu_wx(vuint16m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m1_m))) +vuint16m1_t vwsubu_wx(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m2))) +vuint16m2_t vwsubu_wx(vuint16m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m2_m))) +vuint16m2_t vwsubu_wx(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m4))) +vuint16m4_t vwsubu_wx(vuint16m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m4_m))) +vuint16m4_t vwsubu_wx(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m8))) +vuint16m8_t vwsubu_wx(vuint16m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m8_m))) +vuint16m8_t vwsubu_wx(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32mf2))) +vuint32mf2_t vwsubu_wx(vuint32mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32mf2_m))) +vuint32mf2_t vwsubu_wx(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m1))) +vuint32m1_t vwsubu_wx(vuint32m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m1_m))) +vuint32m1_t vwsubu_wx(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m2))) +vuint32m2_t vwsubu_wx(vuint32m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m2_m))) +vuint32m2_t vwsubu_wx(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m4))) +vuint32m4_t vwsubu_wx(vuint32m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m4_m))) +vuint32m4_t vwsubu_wx(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m8))) +vuint32m8_t vwsubu_wx(vuint32m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m8_m))) +vuint32m8_t vwsubu_wx(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m1))) +vuint64m1_t vwsubu_wx(vuint64m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m1_m))) +vuint64m1_t vwsubu_wx(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m2))) +vuint64m2_t vwsubu_wx(vuint64m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m2_m))) +vuint64m2_t vwsubu_wx(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m4))) +vuint64m4_t vwsubu_wx(vuint64m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m4_m))) +vuint64m4_t vwsubu_wx(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m8))) +vuint64m8_t vwsubu_wx(vuint64m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m8_m))) +vuint64m8_t vwsubu_wx(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16mf4))) +vint16mf4_t vwadd_wv(vint16mf4_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16mf4_m))) +vint16mf4_t vwadd_wv(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16mf2))) +vint16mf2_t vwadd_wv(vint16mf2_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16mf2_m))) +vint16mf2_t vwadd_wv(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m1))) +vint16m1_t vwadd_wv(vint16m1_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m1_m))) +vint16m1_t vwadd_wv(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m2))) +vint16m2_t vwadd_wv(vint16m2_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m2_m))) +vint16m2_t vwadd_wv(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m4))) +vint16m4_t vwadd_wv(vint16m4_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m4_m))) +vint16m4_t vwadd_wv(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m8))) +vint16m8_t vwadd_wv(vint16m8_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m8_m))) +vint16m8_t vwadd_wv(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32mf2))) +vint32mf2_t vwadd_wv(vint32mf2_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32mf2_m))) +vint32mf2_t vwadd_wv(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m1))) +vint32m1_t vwadd_wv(vint32m1_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m1_m))) +vint32m1_t vwadd_wv(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m2))) +vint32m2_t vwadd_wv(vint32m2_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m2_m))) +vint32m2_t vwadd_wv(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m4))) +vint32m4_t vwadd_wv(vint32m4_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m4_m))) +vint32m4_t vwadd_wv(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m8))) +vint32m8_t vwadd_wv(vint32m8_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m8_m))) +vint32m8_t vwadd_wv(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m1))) +vint64m1_t vwadd_wv(vint64m1_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m1_m))) +vint64m1_t vwadd_wv(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m2))) +vint64m2_t vwadd_wv(vint64m2_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m2_m))) +vint64m2_t vwadd_wv(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m4))) +vint64m4_t vwadd_wv(vint64m4_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m4_m))) +vint64m4_t vwadd_wv(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m8))) +vint64m8_t vwadd_wv(vint64m8_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m8_m))) +vint64m8_t vwadd_wv(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16mf4))) +vint16mf4_t vwadd_wx(vint16mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16mf4_m))) +vint16mf4_t vwadd_wx(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16mf2))) +vint16mf2_t vwadd_wx(vint16mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16mf2_m))) +vint16mf2_t vwadd_wx(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m1))) +vint16m1_t vwadd_wx(vint16m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m1_m))) +vint16m1_t vwadd_wx(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m2))) +vint16m2_t vwadd_wx(vint16m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m2_m))) +vint16m2_t vwadd_wx(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m4))) +vint16m4_t vwadd_wx(vint16m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m4_m))) +vint16m4_t vwadd_wx(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m8))) +vint16m8_t vwadd_wx(vint16m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m8_m))) +vint16m8_t vwadd_wx(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32mf2))) +vint32mf2_t vwadd_wx(vint32mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32mf2_m))) +vint32mf2_t vwadd_wx(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m1))) +vint32m1_t vwadd_wx(vint32m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m1_m))) +vint32m1_t vwadd_wx(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m2))) +vint32m2_t vwadd_wx(vint32m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m2_m))) +vint32m2_t vwadd_wx(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m4))) +vint32m4_t vwadd_wx(vint32m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m4_m))) +vint32m4_t vwadd_wx(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m8))) +vint32m8_t vwadd_wx(vint32m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m8_m))) +vint32m8_t vwadd_wx(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m1))) +vint64m1_t vwadd_wx(vint64m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m1_m))) +vint64m1_t vwadd_wx(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m2))) +vint64m2_t vwadd_wx(vint64m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m2_m))) +vint64m2_t vwadd_wx(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m4))) +vint64m4_t vwadd_wx(vint64m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m4_m))) +vint64m4_t vwadd_wx(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m8))) +vint64m8_t vwadd_wx(vint64m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m8_m))) +vint64m8_t vwadd_wx(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16mf4))) +vint16mf4_t vwsub_wv(vint16mf4_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16mf4_m))) +vint16mf4_t vwsub_wv(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16mf2))) +vint16mf2_t vwsub_wv(vint16mf2_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16mf2_m))) +vint16mf2_t vwsub_wv(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m1))) +vint16m1_t vwsub_wv(vint16m1_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m1_m))) +vint16m1_t vwsub_wv(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m2))) +vint16m2_t vwsub_wv(vint16m2_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m2_m))) +vint16m2_t vwsub_wv(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m4))) +vint16m4_t vwsub_wv(vint16m4_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m4_m))) +vint16m4_t vwsub_wv(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m8))) +vint16m8_t vwsub_wv(vint16m8_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m8_m))) +vint16m8_t vwsub_wv(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32mf2))) +vint32mf2_t vwsub_wv(vint32mf2_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32mf2_m))) +vint32mf2_t vwsub_wv(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m1))) +vint32m1_t vwsub_wv(vint32m1_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m1_m))) +vint32m1_t vwsub_wv(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m2))) +vint32m2_t vwsub_wv(vint32m2_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m2_m))) +vint32m2_t vwsub_wv(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m4))) +vint32m4_t vwsub_wv(vint32m4_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m4_m))) +vint32m4_t vwsub_wv(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m8))) +vint32m8_t vwsub_wv(vint32m8_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m8_m))) +vint32m8_t vwsub_wv(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m1))) +vint64m1_t vwsub_wv(vint64m1_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m1_m))) +vint64m1_t vwsub_wv(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m2))) +vint64m2_t vwsub_wv(vint64m2_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m2_m))) +vint64m2_t vwsub_wv(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m4))) +vint64m4_t vwsub_wv(vint64m4_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m4_m))) +vint64m4_t vwsub_wv(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m8))) +vint64m8_t vwsub_wv(vint64m8_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m8_m))) +vint64m8_t vwsub_wv(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16mf4))) +vint16mf4_t vwsub_wx(vint16mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16mf4_m))) +vint16mf4_t vwsub_wx(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16mf2))) +vint16mf2_t vwsub_wx(vint16mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16mf2_m))) +vint16mf2_t vwsub_wx(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m1))) +vint16m1_t vwsub_wx(vint16m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m1_m))) +vint16m1_t vwsub_wx(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m2))) +vint16m2_t vwsub_wx(vint16m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m2_m))) +vint16m2_t vwsub_wx(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m4))) +vint16m4_t vwsub_wx(vint16m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m4_m))) +vint16m4_t vwsub_wx(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m8))) +vint16m8_t vwsub_wx(vint16m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m8_m))) +vint16m8_t vwsub_wx(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32mf2))) +vint32mf2_t vwsub_wx(vint32mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32mf2_m))) +vint32mf2_t vwsub_wx(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m1))) +vint32m1_t vwsub_wx(vint32m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m1_m))) +vint32m1_t vwsub_wx(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m2))) +vint32m2_t vwsub_wx(vint32m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m2_m))) +vint32m2_t vwsub_wx(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m4))) +vint32m4_t vwsub_wx(vint32m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m4_m))) +vint32m4_t vwsub_wx(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m8))) +vint32m8_t vwsub_wx(vint32m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m8_m))) +vint32m8_t vwsub_wx(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m1))) +vint64m1_t vwsub_wx(vint64m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m1_m))) +vint64m1_t vwsub_wx(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m2))) +vint64m2_t vwsub_wx(vint64m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m2_m))) +vint64m2_t vwsub_wx(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m4))) +vint64m4_t vwsub_wx(vint64m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m4_m))) +vint64m4_t vwsub_wx(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m8))) +vint64m8_t vwsub_wx(vint64m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m8_m))) +vint64m8_t vwsub_wx(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m2))) +vint16m2_t vwcvt_x(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m2_m))) +vint16m2_t vwcvt_x(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m4))) +vint16m4_t vwcvt_x(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m4_m))) +vint16m4_t vwcvt_x(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m8))) +vint16m8_t vwcvt_x(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m8_m))) +vint16m8_t vwcvt_x(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m1))) +vint16m1_t vwcvt_x(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m1_m))) +vint16m1_t vwcvt_x(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16mf2))) +vint16mf2_t vwcvt_x(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16mf2_m))) +vint16mf2_t vwcvt_x(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16mf4))) +vint16mf4_t vwcvt_x(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16mf4_m))) +vint16mf4_t vwcvt_x(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m2))) +vint32m2_t vwcvt_x(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m2_m))) +vint32m2_t vwcvt_x(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m4))) +vint32m4_t vwcvt_x(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m4_m))) +vint32m4_t vwcvt_x(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m8))) +vint32m8_t vwcvt_x(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m8_m))) +vint32m8_t vwcvt_x(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m1))) +vint32m1_t vwcvt_x(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m1_m))) +vint32m1_t vwcvt_x(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32mf2))) +vint32mf2_t vwcvt_x(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32mf2_m))) +vint32mf2_t vwcvt_x(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m2))) +vint64m2_t vwcvt_x(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m2_m))) +vint64m2_t vwcvt_x(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m4))) +vint64m4_t vwcvt_x(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m4_m))) +vint64m4_t vwcvt_x(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m8))) +vint64m8_t vwcvt_x(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m8_m))) +vint64m8_t vwcvt_x(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m1))) +vint64m1_t vwcvt_x(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m1_m))) +vint64m1_t vwcvt_x(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8m1))) +void vsuxei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8m1_m))) +void vsuxei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8mf2))) +void vsuxei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8mf2_m))) +void vsuxei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8mf4))) +void vsuxei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8mf4_m))) +void vsuxei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8mf8))) +void vsuxei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8mf8_m))) +void vsuxei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8m1))) +vint8m1_t vadc(vint8m1_t op0, vint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8m2))) +vint8m2_t vadc(vint8m2_t op0, vint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8m4))) +vint8m4_t vadc(vint8m4_t op0, vint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8m8))) +vint8m8_t vadc(vint8m8_t op0, vint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8mf2))) +vint8mf2_t vadc(vint8mf2_t op0, vint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8mf4))) +vint8mf4_t vadc(vint8mf4_t op0, vint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8mf8))) +vint8mf8_t vadc(vint8mf8_t op0, vint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i16m1))) +vint16m1_t vadc(vint16m1_t op0, vint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i16m2))) +vint16m2_t vadc(vint16m2_t op0, vint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i16m4))) +vint16m4_t vadc(vint16m4_t op0, vint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i16m8))) +vint16m8_t vadc(vint16m8_t op0, vint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i16mf2))) +vint16mf2_t vadc(vint16mf2_t op0, vint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i16mf4))) +vint16mf4_t vadc(vint16mf4_t op0, vint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i32m1))) +vint32m1_t vadc(vint32m1_t op0, vint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i32m2))) +vint32m2_t vadc(vint32m2_t op0, vint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i32m4))) +vint32m4_t vadc(vint32m4_t op0, vint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i32m8))) +vint32m8_t vadc(vint32m8_t op0, vint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i32mf2))) +vint32mf2_t vadc(vint32mf2_t op0, vint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i64m1))) +vint64m1_t vadc(vint64m1_t op0, vint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i64m2))) +vint64m2_t vadc(vint64m2_t op0, vint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i64m4))) +vint64m4_t vadc(vint64m4_t op0, vint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i64m8))) +vint64m8_t vadc(vint64m8_t op0, vint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8m1))) +vint8m1_t vadc(vint8m1_t op0, int8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8m2))) +vint8m2_t vadc(vint8m2_t op0, int8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8m4))) +vint8m4_t vadc(vint8m4_t op0, int8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8m8))) +vint8m8_t vadc(vint8m8_t op0, int8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8mf2))) +vint8mf2_t vadc(vint8mf2_t op0, int8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8mf4))) +vint8mf4_t vadc(vint8mf4_t op0, int8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8mf8))) +vint8mf8_t vadc(vint8mf8_t op0, int8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i16m1))) +vint16m1_t vadc(vint16m1_t op0, int16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i16m2))) +vint16m2_t vadc(vint16m2_t op0, int16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i16m4))) +vint16m4_t vadc(vint16m4_t op0, int16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i16m8))) +vint16m8_t vadc(vint16m8_t op0, int16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i16mf2))) +vint16mf2_t vadc(vint16mf2_t op0, int16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i16mf4))) +vint16mf4_t vadc(vint16mf4_t op0, int16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i32m1))) +vint32m1_t vadc(vint32m1_t op0, int32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i32m2))) +vint32m2_t vadc(vint32m2_t op0, int32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i32m4))) +vint32m4_t vadc(vint32m4_t op0, int32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i32m8))) +vint32m8_t vadc(vint32m8_t op0, int32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i32mf2))) +vint32mf2_t vadc(vint32mf2_t op0, int32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i64m1))) +vint64m1_t vadc(vint64m1_t op0, int64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i64m2))) +vint64m2_t vadc(vint64m2_t op0, int64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i64m4))) +vint64m4_t vadc(vint64m4_t op0, int64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i64m8))) +vint64m8_t vadc(vint64m8_t op0, int64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8m1))) +vuint8m1_t vadc(vuint8m1_t op0, vuint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8m2))) +vuint8m2_t vadc(vuint8m2_t op0, vuint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8m4))) +vuint8m4_t vadc(vuint8m4_t op0, vuint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8m8))) +vuint8m8_t vadc(vuint8m8_t op0, vuint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8mf2))) +vuint8mf2_t vadc(vuint8mf2_t op0, vuint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8mf4))) +vuint8mf4_t vadc(vuint8mf4_t op0, vuint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8mf8))) +vuint8mf8_t vadc(vuint8mf8_t op0, vuint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u16m1))) +vuint16m1_t vadc(vuint16m1_t op0, vuint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u16m2))) +vuint16m2_t vadc(vuint16m2_t op0, vuint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u16m4))) +vuint16m4_t vadc(vuint16m4_t op0, vuint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u16m8))) +vuint16m8_t vadc(vuint16m8_t op0, vuint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u16mf2))) +vuint16mf2_t vadc(vuint16mf2_t op0, vuint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u16mf4))) +vuint16mf4_t vadc(vuint16mf4_t op0, vuint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u32m1))) +vuint32m1_t vadc(vuint32m1_t op0, vuint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u32m2))) +vuint32m2_t vadc(vuint32m2_t op0, vuint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u32m4))) +vuint32m4_t vadc(vuint32m4_t op0, vuint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u32m8))) +vuint32m8_t vadc(vuint32m8_t op0, vuint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u32mf2))) +vuint32mf2_t vadc(vuint32mf2_t op0, vuint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u64m1))) +vuint64m1_t vadc(vuint64m1_t op0, vuint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u64m2))) +vuint64m2_t vadc(vuint64m2_t op0, vuint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u64m4))) +vuint64m4_t vadc(vuint64m4_t op0, vuint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u64m8))) +vuint64m8_t vadc(vuint64m8_t op0, vuint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8m1))) +vuint8m1_t vadc(vuint8m1_t op0, uint8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8m2))) +vuint8m2_t vadc(vuint8m2_t op0, uint8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8m4))) +vuint8m4_t vadc(vuint8m4_t op0, uint8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8m8))) +vuint8m8_t vadc(vuint8m8_t op0, uint8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8mf2))) +vuint8mf2_t vadc(vuint8mf2_t op0, uint8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8mf4))) +vuint8mf4_t vadc(vuint8mf4_t op0, uint8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8mf8))) +vuint8mf8_t vadc(vuint8mf8_t op0, uint8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u16m1))) +vuint16m1_t vadc(vuint16m1_t op0, uint16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u16m2))) +vuint16m2_t vadc(vuint16m2_t op0, uint16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u16m4))) +vuint16m4_t vadc(vuint16m4_t op0, uint16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u16m8))) +vuint16m8_t vadc(vuint16m8_t op0, uint16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u16mf2))) +vuint16mf2_t vadc(vuint16mf2_t op0, uint16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u16mf4))) +vuint16mf4_t vadc(vuint16mf4_t op0, uint16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u32m1))) +vuint32m1_t vadc(vuint32m1_t op0, uint32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u32m2))) +vuint32m2_t vadc(vuint32m2_t op0, uint32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u32m4))) +vuint32m4_t vadc(vuint32m4_t op0, uint32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u32m8))) +vuint32m8_t vadc(vuint32m8_t op0, uint32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u32mf2))) +vuint32mf2_t vadc(vuint32mf2_t op0, uint32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u64m1))) +vuint64m1_t vadc(vuint64m1_t op0, uint64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u64m2))) +vuint64m2_t vadc(vuint64m2_t op0, uint64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u64m4))) +vuint64m4_t vadc(vuint64m4_t op0, uint64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u64m8))) +vuint64m8_t vadc(vuint64m8_t op0, uint64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8m1_b8))) +vbool8_t vmadc(vint8m1_t op0, vint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8m2_b4))) +vbool4_t vmadc(vint8m2_t op0, vint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8m4_b2))) +vbool2_t vmadc(vint8m4_t op0, vint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8m8_b1))) +vbool1_t vmadc(vint8m8_t op0, vint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8mf2_b16))) +vbool16_t vmadc(vint8mf2_t op0, vint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8mf4_b32))) +vbool32_t vmadc(vint8mf4_t op0, vint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8mf8_b64))) +vbool64_t vmadc(vint8mf8_t op0, vint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i16m1_b16))) +vbool16_t vmadc(vint16m1_t op0, vint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i16m2_b8))) +vbool8_t vmadc(vint16m2_t op0, vint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i16m4_b4))) +vbool4_t vmadc(vint16m4_t op0, vint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i16m8_b2))) +vbool2_t vmadc(vint16m8_t op0, vint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i16mf2_b32))) +vbool32_t vmadc(vint16mf2_t op0, vint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i16mf4_b64))) +vbool64_t vmadc(vint16mf4_t op0, vint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i32m1_b32))) +vbool32_t vmadc(vint32m1_t op0, vint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i32m2_b16))) +vbool16_t vmadc(vint32m2_t op0, vint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i32m4_b8))) +vbool8_t vmadc(vint32m4_t op0, vint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i32m8_b4))) +vbool4_t vmadc(vint32m8_t op0, vint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i32mf2_b64))) +vbool64_t vmadc(vint32mf2_t op0, vint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i64m1_b64))) +vbool64_t vmadc(vint64m1_t op0, vint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i64m2_b32))) +vbool32_t vmadc(vint64m2_t op0, vint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i64m4_b16))) +vbool16_t vmadc(vint64m4_t op0, vint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i64m8_b8))) +vbool8_t vmadc(vint64m8_t op0, vint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8m1_b8))) +vbool8_t vmadc(vint8m1_t op0, int8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8m2_b4))) +vbool4_t vmadc(vint8m2_t op0, int8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8m4_b2))) +vbool2_t vmadc(vint8m4_t op0, int8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8m8_b1))) +vbool1_t vmadc(vint8m8_t op0, int8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8mf2_b16))) +vbool16_t vmadc(vint8mf2_t op0, int8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8mf4_b32))) +vbool32_t vmadc(vint8mf4_t op0, int8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8mf8_b64))) +vbool64_t vmadc(vint8mf8_t op0, int8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i16m1_b16))) +vbool16_t vmadc(vint16m1_t op0, int16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i16m2_b8))) +vbool8_t vmadc(vint16m2_t op0, int16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i16m4_b4))) +vbool4_t vmadc(vint16m4_t op0, int16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i16m8_b2))) +vbool2_t vmadc(vint16m8_t op0, int16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i16mf2_b32))) +vbool32_t vmadc(vint16mf2_t op0, int16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i16mf4_b64))) +vbool64_t vmadc(vint16mf4_t op0, int16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i32m1_b32))) +vbool32_t vmadc(vint32m1_t op0, int32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i32m2_b16))) +vbool16_t vmadc(vint32m2_t op0, int32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i32m4_b8))) +vbool8_t vmadc(vint32m4_t op0, int32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i32m8_b4))) +vbool4_t vmadc(vint32m8_t op0, int32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i32mf2_b64))) +vbool64_t vmadc(vint32mf2_t op0, int32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i64m1_b64))) +vbool64_t vmadc(vint64m1_t op0, int64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i64m2_b32))) +vbool32_t vmadc(vint64m2_t op0, int64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i64m4_b16))) +vbool16_t vmadc(vint64m4_t op0, int64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i64m8_b8))) +vbool8_t vmadc(vint64m8_t op0, int64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8m1_b8))) +vbool8_t vmadc(vuint8m1_t op0, vuint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8m2_b4))) +vbool4_t vmadc(vuint8m2_t op0, vuint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8m4_b2))) +vbool2_t vmadc(vuint8m4_t op0, vuint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8m8_b1))) +vbool1_t vmadc(vuint8m8_t op0, vuint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8mf2_b16))) +vbool16_t vmadc(vuint8mf2_t op0, vuint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8mf4_b32))) +vbool32_t vmadc(vuint8mf4_t op0, vuint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8mf8_b64))) +vbool64_t vmadc(vuint8mf8_t op0, vuint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u16m1_b16))) +vbool16_t vmadc(vuint16m1_t op0, vuint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u16m2_b8))) +vbool8_t vmadc(vuint16m2_t op0, vuint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u16m4_b4))) +vbool4_t vmadc(vuint16m4_t op0, vuint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u16m8_b2))) +vbool2_t vmadc(vuint16m8_t op0, vuint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u16mf2_b32))) +vbool32_t vmadc(vuint16mf2_t op0, vuint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u16mf4_b64))) +vbool64_t vmadc(vuint16mf4_t op0, vuint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u32m1_b32))) +vbool32_t vmadc(vuint32m1_t op0, vuint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u32m2_b16))) +vbool16_t vmadc(vuint32m2_t op0, vuint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u32m4_b8))) +vbool8_t vmadc(vuint32m4_t op0, vuint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u32m8_b4))) +vbool4_t vmadc(vuint32m8_t op0, vuint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u32mf2_b64))) +vbool64_t vmadc(vuint32mf2_t op0, vuint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u64m1_b64))) +vbool64_t vmadc(vuint64m1_t op0, vuint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u64m2_b32))) +vbool32_t vmadc(vuint64m2_t op0, vuint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u64m4_b16))) +vbool16_t vmadc(vuint64m4_t op0, vuint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u64m8_b8))) +vbool8_t vmadc(vuint64m8_t op0, vuint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8m1_b8))) +vbool8_t vmadc(vuint8m1_t op0, uint8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8m2_b4))) +vbool4_t vmadc(vuint8m2_t op0, uint8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8m4_b2))) +vbool2_t vmadc(vuint8m4_t op0, uint8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8m8_b1))) +vbool1_t vmadc(vuint8m8_t op0, uint8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8mf2_b16))) +vbool16_t vmadc(vuint8mf2_t op0, uint8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8mf4_b32))) +vbool32_t vmadc(vuint8mf4_t op0, uint8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8mf8_b64))) +vbool64_t vmadc(vuint8mf8_t op0, uint8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u16m1_b16))) +vbool16_t vmadc(vuint16m1_t op0, uint16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u16m2_b8))) +vbool8_t vmadc(vuint16m2_t op0, uint16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u16m4_b4))) +vbool4_t vmadc(vuint16m4_t op0, uint16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u16m8_b2))) +vbool2_t vmadc(vuint16m8_t op0, uint16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u16mf2_b32))) +vbool32_t vmadc(vuint16mf2_t op0, uint16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u16mf4_b64))) +vbool64_t vmadc(vuint16mf4_t op0, uint16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u32m1_b32))) +vbool32_t vmadc(vuint32m1_t op0, uint32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u32m2_b16))) +vbool16_t vmadc(vuint32m2_t op0, uint32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u32m4_b8))) +vbool8_t vmadc(vuint32m4_t op0, uint32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u32m8_b4))) +vbool4_t vmadc(vuint32m8_t op0, uint32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u32mf2_b64))) +vbool64_t vmadc(vuint32mf2_t op0, uint32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u64m1_b64))) +vbool64_t vmadc(vuint64m1_t op0, uint64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u64m2_b32))) +vbool32_t vmadc(vuint64m2_t op0, uint64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u64m4_b16))) +vbool16_t vmadc(vuint64m4_t op0, uint64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u64m8_b8))) +vbool8_t vmadc(vuint64m8_t op0, uint64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8m1_b8))) +vbool8_t vmadc(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8m2_b4))) +vbool4_t vmadc(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8m4_b2))) +vbool2_t vmadc(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8m8_b1))) +vbool1_t vmadc(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8mf2_b16))) +vbool16_t vmadc(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8mf4_b32))) +vbool32_t vmadc(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8mf8_b64))) +vbool64_t vmadc(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i16m1_b16))) +vbool16_t vmadc(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i16m2_b8))) +vbool8_t vmadc(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i16m4_b4))) +vbool4_t vmadc(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i16m8_b2))) +vbool2_t vmadc(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i16mf2_b32))) +vbool32_t vmadc(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i16mf4_b64))) +vbool64_t vmadc(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i32m1_b32))) +vbool32_t vmadc(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i32m2_b16))) +vbool16_t vmadc(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i32m4_b8))) +vbool8_t vmadc(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i32m8_b4))) +vbool4_t vmadc(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i32mf2_b64))) +vbool64_t vmadc(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i64m1_b64))) +vbool64_t vmadc(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i64m2_b32))) +vbool32_t vmadc(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i64m4_b16))) +vbool16_t vmadc(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i64m8_b8))) +vbool8_t vmadc(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8m1_b8))) +vbool8_t vmadc(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8m2_b4))) +vbool4_t vmadc(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8m4_b2))) +vbool2_t vmadc(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8m8_b1))) +vbool1_t vmadc(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8mf2_b16))) +vbool16_t vmadc(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8mf4_b32))) +vbool32_t vmadc(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8mf8_b64))) +vbool64_t vmadc(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i16m1_b16))) +vbool16_t vmadc(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i16m2_b8))) +vbool8_t vmadc(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i16m4_b4))) +vbool4_t vmadc(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i16m8_b2))) +vbool2_t vmadc(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i16mf2_b32))) +vbool32_t vmadc(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i16mf4_b64))) +vbool64_t vmadc(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i32m1_b32))) +vbool32_t vmadc(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i32m2_b16))) +vbool16_t vmadc(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i32m4_b8))) +vbool8_t vmadc(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i32m8_b4))) +vbool4_t vmadc(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i32mf2_b64))) +vbool64_t vmadc(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i64m1_b64))) +vbool64_t vmadc(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i64m2_b32))) +vbool32_t vmadc(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i64m4_b16))) +vbool16_t vmadc(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i64m8_b8))) +vbool8_t vmadc(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m1))) +void vsuxei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m1_m))) +void vsuxei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m2))) +void vsuxei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m2_m))) +void vsuxei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m4))) +void vsuxei8(int16_t * op0, vuint8m2_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m4_m))) +void vsuxei8(vbool4_t op0, int16_t * op1, vuint8m2_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m8))) +void vsuxei8(int16_t * op0, vuint8m4_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m8_m))) +void vsuxei8(vbool2_t op0, int16_t * op1, vuint8m4_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16mf2))) +void vsuxei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16mf2_m))) +void vsuxei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16mf4))) +void vsuxei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16mf4_m))) +void vsuxei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8m1_b8))) +vbool8_t vmadc(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8m2_b4))) +vbool4_t vmadc(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8m4_b2))) +vbool2_t vmadc(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8m8_b1))) +vbool1_t vmadc(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8mf2_b16))) +vbool16_t vmadc(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8mf4_b32))) +vbool32_t vmadc(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8mf8_b64))) +vbool64_t vmadc(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u16m1_b16))) +vbool16_t vmadc(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u16m2_b8))) +vbool8_t vmadc(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u16m4_b4))) +vbool4_t vmadc(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u16m8_b2))) +vbool2_t vmadc(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u16mf2_b32))) +vbool32_t vmadc(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u16mf4_b64))) +vbool64_t vmadc(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u32m1_b32))) +vbool32_t vmadc(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u32m2_b16))) +vbool16_t vmadc(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u32m4_b8))) +vbool8_t vmadc(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u32m8_b4))) +vbool4_t vmadc(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u32mf2_b64))) +vbool64_t vmadc(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u64m1_b64))) +vbool64_t vmadc(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u64m2_b32))) +vbool32_t vmadc(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u64m4_b16))) +vbool16_t vmadc(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u64m8_b8))) +vbool8_t vmadc(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8m1_b8))) +vbool8_t vmadc(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8m2_b4))) +vbool4_t vmadc(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8m4_b2))) +vbool2_t vmadc(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8m8_b1))) +vbool1_t vmadc(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8mf2_b16))) +vbool16_t vmadc(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8mf4_b32))) +vbool32_t vmadc(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8mf8_b64))) +vbool64_t vmadc(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u16m1_b16))) +vbool16_t vmadc(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u16m2_b8))) +vbool8_t vmadc(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u16m4_b4))) +vbool4_t vmadc(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u16m8_b2))) +vbool2_t vmadc(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u16mf2_b32))) +vbool32_t vmadc(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u16mf4_b64))) +vbool64_t vmadc(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u32m1_b32))) +vbool32_t vmadc(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u32m2_b16))) +vbool16_t vmadc(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u32m4_b8))) +vbool8_t vmadc(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u32m8_b4))) +vbool4_t vmadc(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u32mf2_b64))) +vbool64_t vmadc(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u64m1_b64))) +vbool64_t vmadc(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u64m2_b32))) +vbool32_t vmadc(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u64m4_b16))) +vbool16_t vmadc(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u64m8_b8))) +vbool8_t vmadc(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8m1))) +vint8m1_t vsbc(vint8m1_t op0, vint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8m2))) +vint8m2_t vsbc(vint8m2_t op0, vint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8m4))) +vint8m4_t vsbc(vint8m4_t op0, vint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8m8))) +vint8m8_t vsbc(vint8m8_t op0, vint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8mf2))) +vint8mf2_t vsbc(vint8mf2_t op0, vint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8mf4))) +vint8mf4_t vsbc(vint8mf4_t op0, vint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8mf8))) +vint8mf8_t vsbc(vint8mf8_t op0, vint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i16m1))) +vint16m1_t vsbc(vint16m1_t op0, vint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i16m2))) +vint16m2_t vsbc(vint16m2_t op0, vint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i16m4))) +vint16m4_t vsbc(vint16m4_t op0, vint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i16m8))) +vint16m8_t vsbc(vint16m8_t op0, vint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i16mf2))) +vint16mf2_t vsbc(vint16mf2_t op0, vint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i16mf4))) +vint16mf4_t vsbc(vint16mf4_t op0, vint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i32m1))) +vint32m1_t vsbc(vint32m1_t op0, vint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i32m2))) +vint32m2_t vsbc(vint32m2_t op0, vint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i32m4))) +vint32m4_t vsbc(vint32m4_t op0, vint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i32m8))) +vint32m8_t vsbc(vint32m8_t op0, vint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i32mf2))) +vint32mf2_t vsbc(vint32mf2_t op0, vint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i64m1))) +vint64m1_t vsbc(vint64m1_t op0, vint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i64m2))) +vint64m2_t vsbc(vint64m2_t op0, vint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i64m4))) +vint64m4_t vsbc(vint64m4_t op0, vint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i64m8))) +vint64m8_t vsbc(vint64m8_t op0, vint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8m1))) +vint8m1_t vsbc(vint8m1_t op0, int8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8m2))) +vint8m2_t vsbc(vint8m2_t op0, int8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8m4))) +vint8m4_t vsbc(vint8m4_t op0, int8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8m8))) +vint8m8_t vsbc(vint8m8_t op0, int8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8mf2))) +vint8mf2_t vsbc(vint8mf2_t op0, int8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8mf4))) +vint8mf4_t vsbc(vint8mf4_t op0, int8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8mf8))) +vint8mf8_t vsbc(vint8mf8_t op0, int8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i16m1))) +vint16m1_t vsbc(vint16m1_t op0, int16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i16m2))) +vint16m2_t vsbc(vint16m2_t op0, int16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i16m4))) +vint16m4_t vsbc(vint16m4_t op0, int16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i16m8))) +vint16m8_t vsbc(vint16m8_t op0, int16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i16mf2))) +vint16mf2_t vsbc(vint16mf2_t op0, int16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i16mf4))) +vint16mf4_t vsbc(vint16mf4_t op0, int16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i32m1))) +vint32m1_t vsbc(vint32m1_t op0, int32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i32m2))) +vint32m2_t vsbc(vint32m2_t op0, int32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i32m4))) +vint32m4_t vsbc(vint32m4_t op0, int32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i32m8))) +vint32m8_t vsbc(vint32m8_t op0, int32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i32mf2))) +vint32mf2_t vsbc(vint32mf2_t op0, int32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i64m1))) +vint64m1_t vsbc(vint64m1_t op0, int64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i64m2))) +vint64m2_t vsbc(vint64m2_t op0, int64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i64m4))) +vint64m4_t vsbc(vint64m4_t op0, int64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i64m8))) +vint64m8_t vsbc(vint64m8_t op0, int64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8m1))) +vuint8m1_t vsbc(vuint8m1_t op0, vuint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8m2))) +vuint8m2_t vsbc(vuint8m2_t op0, vuint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8m4))) +vuint8m4_t vsbc(vuint8m4_t op0, vuint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8m8))) +vuint8m8_t vsbc(vuint8m8_t op0, vuint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8mf2))) +vuint8mf2_t vsbc(vuint8mf2_t op0, vuint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8mf4))) +vuint8mf4_t vsbc(vuint8mf4_t op0, vuint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8mf8))) +vuint8mf8_t vsbc(vuint8mf8_t op0, vuint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u16m1))) +vuint16m1_t vsbc(vuint16m1_t op0, vuint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u16m2))) +vuint16m2_t vsbc(vuint16m2_t op0, vuint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u16m4))) +vuint16m4_t vsbc(vuint16m4_t op0, vuint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u16m8))) +vuint16m8_t vsbc(vuint16m8_t op0, vuint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u16mf2))) +vuint16mf2_t vsbc(vuint16mf2_t op0, vuint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u16mf4))) +vuint16mf4_t vsbc(vuint16mf4_t op0, vuint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u32m1))) +vuint32m1_t vsbc(vuint32m1_t op0, vuint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u32m2))) +vuint32m2_t vsbc(vuint32m2_t op0, vuint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u32m4))) +vuint32m4_t vsbc(vuint32m4_t op0, vuint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u32m8))) +vuint32m8_t vsbc(vuint32m8_t op0, vuint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u32mf2))) +vuint32mf2_t vsbc(vuint32mf2_t op0, vuint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u64m1))) +vuint64m1_t vsbc(vuint64m1_t op0, vuint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u64m2))) +vuint64m2_t vsbc(vuint64m2_t op0, vuint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u64m4))) +vuint64m4_t vsbc(vuint64m4_t op0, vuint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u64m8))) +vuint64m8_t vsbc(vuint64m8_t op0, vuint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8m1))) +vuint8m1_t vsbc(vuint8m1_t op0, uint8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8m2))) +vuint8m2_t vsbc(vuint8m2_t op0, uint8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8m4))) +vuint8m4_t vsbc(vuint8m4_t op0, uint8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8m8))) +vuint8m8_t vsbc(vuint8m8_t op0, uint8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8mf2))) +vuint8mf2_t vsbc(vuint8mf2_t op0, uint8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8mf4))) +vuint8mf4_t vsbc(vuint8mf4_t op0, uint8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8mf8))) +vuint8mf8_t vsbc(vuint8mf8_t op0, uint8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u16m1))) +vuint16m1_t vsbc(vuint16m1_t op0, uint16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u16m2))) +vuint16m2_t vsbc(vuint16m2_t op0, uint16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u16m4))) +vuint16m4_t vsbc(vuint16m4_t op0, uint16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u16m8))) +vuint16m8_t vsbc(vuint16m8_t op0, uint16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u16mf2))) +vuint16mf2_t vsbc(vuint16mf2_t op0, uint16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u16mf4))) +vuint16mf4_t vsbc(vuint16mf4_t op0, uint16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u32m1))) +vuint32m1_t vsbc(vuint32m1_t op0, uint32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u32m2))) +vuint32m2_t vsbc(vuint32m2_t op0, uint32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u32m4))) +vuint32m4_t vsbc(vuint32m4_t op0, uint32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u32m8))) +vuint32m8_t vsbc(vuint32m8_t op0, uint32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u32mf2))) +vuint32mf2_t vsbc(vuint32mf2_t op0, uint32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u64m1))) +vuint64m1_t vsbc(vuint64m1_t op0, uint64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u64m2))) +vuint64m2_t vsbc(vuint64m2_t op0, uint64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u64m4))) +vuint64m4_t vsbc(vuint64m4_t op0, uint64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u64m8))) +vuint64m8_t vsbc(vuint64m8_t op0, uint64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8m1_b8))) +vbool8_t vmsbc(vint8m1_t op0, vint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8m2_b4))) +vbool4_t vmsbc(vint8m2_t op0, vint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8m4_b2))) +vbool2_t vmsbc(vint8m4_t op0, vint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8m8_b1))) +vbool1_t vmsbc(vint8m8_t op0, vint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8mf2_b16))) +vbool16_t vmsbc(vint8mf2_t op0, vint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8mf4_b32))) +vbool32_t vmsbc(vint8mf4_t op0, vint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8mf8_b64))) +vbool64_t vmsbc(vint8mf8_t op0, vint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i16m1_b16))) +vbool16_t vmsbc(vint16m1_t op0, vint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i16m2_b8))) +vbool8_t vmsbc(vint16m2_t op0, vint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i16m4_b4))) +vbool4_t vmsbc(vint16m4_t op0, vint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i16m8_b2))) +vbool2_t vmsbc(vint16m8_t op0, vint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i16mf2_b32))) +vbool32_t vmsbc(vint16mf2_t op0, vint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i16mf4_b64))) +vbool64_t vmsbc(vint16mf4_t op0, vint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i32m1_b32))) +vbool32_t vmsbc(vint32m1_t op0, vint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i32m2_b16))) +vbool16_t vmsbc(vint32m2_t op0, vint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i32m4_b8))) +vbool8_t vmsbc(vint32m4_t op0, vint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i32m8_b4))) +vbool4_t vmsbc(vint32m8_t op0, vint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i32mf2_b64))) +vbool64_t vmsbc(vint32mf2_t op0, vint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i64m1_b64))) +vbool64_t vmsbc(vint64m1_t op0, vint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i64m2_b32))) +vbool32_t vmsbc(vint64m2_t op0, vint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i64m4_b16))) +vbool16_t vmsbc(vint64m4_t op0, vint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i64m8_b8))) +vbool8_t vmsbc(vint64m8_t op0, vint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8m1_b8))) +vbool8_t vmsbc(vint8m1_t op0, int8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8m2_b4))) +vbool4_t vmsbc(vint8m2_t op0, int8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8m4_b2))) +vbool2_t vmsbc(vint8m4_t op0, int8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8m8_b1))) +vbool1_t vmsbc(vint8m8_t op0, int8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8mf2_b16))) +vbool16_t vmsbc(vint8mf2_t op0, int8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8mf4_b32))) +vbool32_t vmsbc(vint8mf4_t op0, int8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8mf8_b64))) +vbool64_t vmsbc(vint8mf8_t op0, int8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i16m1_b16))) +vbool16_t vmsbc(vint16m1_t op0, int16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i16m2_b8))) +vbool8_t vmsbc(vint16m2_t op0, int16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i16m4_b4))) +vbool4_t vmsbc(vint16m4_t op0, int16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i16m8_b2))) +vbool2_t vmsbc(vint16m8_t op0, int16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i16mf2_b32))) +vbool32_t vmsbc(vint16mf2_t op0, int16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i16mf4_b64))) +vbool64_t vmsbc(vint16mf4_t op0, int16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i32m1_b32))) +vbool32_t vmsbc(vint32m1_t op0, int32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i32m2_b16))) +vbool16_t vmsbc(vint32m2_t op0, int32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i32m4_b8))) +vbool8_t vmsbc(vint32m4_t op0, int32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i32m8_b4))) +vbool4_t vmsbc(vint32m8_t op0, int32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i32mf2_b64))) +vbool64_t vmsbc(vint32mf2_t op0, int32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i64m1_b64))) +vbool64_t vmsbc(vint64m1_t op0, int64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i64m2_b32))) +vbool32_t vmsbc(vint64m2_t op0, int64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i64m4_b16))) +vbool16_t vmsbc(vint64m4_t op0, int64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i64m8_b8))) +vbool8_t vmsbc(vint64m8_t op0, int64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8m1_b8))) +vbool8_t vmsbc(vuint8m1_t op0, vuint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8m2_b4))) +vbool4_t vmsbc(vuint8m2_t op0, vuint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8m4_b2))) +vbool2_t vmsbc(vuint8m4_t op0, vuint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8m8_b1))) +vbool1_t vmsbc(vuint8m8_t op0, vuint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8mf2_b16))) +vbool16_t vmsbc(vuint8mf2_t op0, vuint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8mf4_b32))) +vbool32_t vmsbc(vuint8mf4_t op0, vuint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8mf8_b64))) +vbool64_t vmsbc(vuint8mf8_t op0, vuint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u16m1_b16))) +vbool16_t vmsbc(vuint16m1_t op0, vuint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u16m2_b8))) +vbool8_t vmsbc(vuint16m2_t op0, vuint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u16m4_b4))) +vbool4_t vmsbc(vuint16m4_t op0, vuint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u16m8_b2))) +vbool2_t vmsbc(vuint16m8_t op0, vuint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u16mf2_b32))) +vbool32_t vmsbc(vuint16mf2_t op0, vuint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u16mf4_b64))) +vbool64_t vmsbc(vuint16mf4_t op0, vuint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u32m1_b32))) +vbool32_t vmsbc(vuint32m1_t op0, vuint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u32m2_b16))) +vbool16_t vmsbc(vuint32m2_t op0, vuint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u32m4_b8))) +vbool8_t vmsbc(vuint32m4_t op0, vuint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u32m8_b4))) +vbool4_t vmsbc(vuint32m8_t op0, vuint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u32mf2_b64))) +vbool64_t vmsbc(vuint32mf2_t op0, vuint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u64m1_b64))) +vbool64_t vmsbc(vuint64m1_t op0, vuint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u64m2_b32))) +vbool32_t vmsbc(vuint64m2_t op0, vuint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u64m4_b16))) +vbool16_t vmsbc(vuint64m4_t op0, vuint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u64m8_b8))) +vbool8_t vmsbc(vuint64m8_t op0, vuint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8m1_b8))) +vbool8_t vmsbc(vuint8m1_t op0, uint8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8m2_b4))) +vbool4_t vmsbc(vuint8m2_t op0, uint8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8m4_b2))) +vbool2_t vmsbc(vuint8m4_t op0, uint8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8m8_b1))) +vbool1_t vmsbc(vuint8m8_t op0, uint8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8mf2_b16))) +vbool16_t vmsbc(vuint8mf2_t op0, uint8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8mf4_b32))) +vbool32_t vmsbc(vuint8mf4_t op0, uint8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8mf8_b64))) +vbool64_t vmsbc(vuint8mf8_t op0, uint8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u16m1_b16))) +vbool16_t vmsbc(vuint16m1_t op0, uint16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u16m2_b8))) +vbool8_t vmsbc(vuint16m2_t op0, uint16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u16m4_b4))) +vbool4_t vmsbc(vuint16m4_t op0, uint16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u16m8_b2))) +vbool2_t vmsbc(vuint16m8_t op0, uint16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u16mf2_b32))) +vbool32_t vmsbc(vuint16mf2_t op0, uint16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u16mf4_b64))) +vbool64_t vmsbc(vuint16mf4_t op0, uint16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u32m1_b32))) +vbool32_t vmsbc(vuint32m1_t op0, uint32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u32m2_b16))) +vbool16_t vmsbc(vuint32m2_t op0, uint32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u32m4_b8))) +vbool8_t vmsbc(vuint32m4_t op0, uint32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u32m8_b4))) +vbool4_t vmsbc(vuint32m8_t op0, uint32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u32mf2_b64))) +vbool64_t vmsbc(vuint32mf2_t op0, uint32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u64m1_b64))) +vbool64_t vmsbc(vuint64m1_t op0, uint64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u64m2_b32))) +vbool32_t vmsbc(vuint64m2_t op0, uint64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u64m4_b16))) +vbool16_t vmsbc(vuint64m4_t op0, uint64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u64m8_b8))) +vbool8_t vmsbc(vuint64m8_t op0, uint64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m1))) +void vsuxei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m1_m))) +void vsuxei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m2))) +void vsuxei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m2_m))) +void vsuxei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m4))) +void vsuxei8(uint16_t * op0, vuint8m2_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m4_m))) +void vsuxei8(vbool4_t op0, uint16_t * op1, vuint8m2_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m8))) +void vsuxei8(uint16_t * op0, vuint8m4_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m8_m))) +void vsuxei8(vbool2_t op0, uint16_t * op1, vuint8m4_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16mf2))) +void vsuxei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16mf2_m))) +void vsuxei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16mf4))) +void vsuxei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16mf4_m))) +void vsuxei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8m1_b8))) +vbool8_t vmsbc(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8m2_b4))) +vbool4_t vmsbc(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8m4_b2))) +vbool2_t vmsbc(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8m8_b1))) +vbool1_t vmsbc(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8mf2_b16))) +vbool16_t vmsbc(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8mf4_b32))) +vbool32_t vmsbc(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8mf8_b64))) +vbool64_t vmsbc(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i16m1_b16))) +vbool16_t vmsbc(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i16m2_b8))) +vbool8_t vmsbc(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i16m4_b4))) +vbool4_t vmsbc(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i16m8_b2))) +vbool2_t vmsbc(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i16mf2_b32))) +vbool32_t vmsbc(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i16mf4_b64))) +vbool64_t vmsbc(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i32m1_b32))) +vbool32_t vmsbc(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i32m2_b16))) +vbool16_t vmsbc(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i32m4_b8))) +vbool8_t vmsbc(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i32m8_b4))) +vbool4_t vmsbc(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i32mf2_b64))) +vbool64_t vmsbc(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i64m1_b64))) +vbool64_t vmsbc(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i64m2_b32))) +vbool32_t vmsbc(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i64m4_b16))) +vbool16_t vmsbc(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i64m8_b8))) +vbool8_t vmsbc(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8m1_b8))) +vbool8_t vmsbc(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8m2_b4))) +vbool4_t vmsbc(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8m4_b2))) +vbool2_t vmsbc(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8m8_b1))) +vbool1_t vmsbc(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8mf2_b16))) +vbool16_t vmsbc(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8mf4_b32))) +vbool32_t vmsbc(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8mf8_b64))) +vbool64_t vmsbc(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i16m1_b16))) +vbool16_t vmsbc(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i16m2_b8))) +vbool8_t vmsbc(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i16m4_b4))) +vbool4_t vmsbc(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i16m8_b2))) +vbool2_t vmsbc(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i16mf2_b32))) +vbool32_t vmsbc(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i16mf4_b64))) +vbool64_t vmsbc(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i32m1_b32))) +vbool32_t vmsbc(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i32m2_b16))) +vbool16_t vmsbc(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i32m4_b8))) +vbool8_t vmsbc(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i32m8_b4))) +vbool4_t vmsbc(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i32mf2_b64))) +vbool64_t vmsbc(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i64m1_b64))) +vbool64_t vmsbc(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i64m2_b32))) +vbool32_t vmsbc(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i64m4_b16))) +vbool16_t vmsbc(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i64m8_b8))) +vbool8_t vmsbc(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8m1_b8))) +vbool8_t vmsbc(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8m2_b4))) +vbool4_t vmsbc(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8m4_b2))) +vbool2_t vmsbc(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8m8_b1))) +vbool1_t vmsbc(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8mf2_b16))) +vbool16_t vmsbc(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8mf4_b32))) +vbool32_t vmsbc(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8mf8_b64))) +vbool64_t vmsbc(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u16m1_b16))) +vbool16_t vmsbc(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u16m2_b8))) +vbool8_t vmsbc(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u16m4_b4))) +vbool4_t vmsbc(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u16m8_b2))) +vbool2_t vmsbc(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u16mf2_b32))) +vbool32_t vmsbc(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u16mf4_b64))) +vbool64_t vmsbc(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u32m1_b32))) +vbool32_t vmsbc(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u32m2_b16))) +vbool16_t vmsbc(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u32m4_b8))) +vbool8_t vmsbc(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u32m8_b4))) +vbool4_t vmsbc(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u32mf2_b64))) +vbool64_t vmsbc(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u64m1_b64))) +vbool64_t vmsbc(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u64m2_b32))) +vbool32_t vmsbc(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u64m4_b16))) +vbool16_t vmsbc(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u64m8_b8))) +vbool8_t vmsbc(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8m1_b8))) +vbool8_t vmsbc(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8m2_b4))) +vbool4_t vmsbc(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8m4_b2))) +vbool2_t vmsbc(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8m8_b1))) +vbool1_t vmsbc(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8mf2_b16))) +vbool16_t vmsbc(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8mf4_b32))) +vbool32_t vmsbc(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8mf8_b64))) +vbool64_t vmsbc(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u16m1_b16))) +vbool16_t vmsbc(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u16m2_b8))) +vbool8_t vmsbc(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u16m4_b4))) +vbool4_t vmsbc(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u16m8_b2))) +vbool2_t vmsbc(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u16mf2_b32))) +vbool32_t vmsbc(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u16mf4_b64))) +vbool64_t vmsbc(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u32m1_b32))) +vbool32_t vmsbc(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u32m2_b16))) +vbool16_t vmsbc(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u32m4_b8))) +vbool8_t vmsbc(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u32m8_b4))) +vbool4_t vmsbc(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u32mf2_b64))) +vbool64_t vmsbc(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u64m1_b64))) +vbool64_t vmsbc(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u64m2_b32))) +vbool32_t vmsbc(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u64m4_b16))) +vbool16_t vmsbc(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u64m8_b8))) +vbool8_t vmsbc(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m1))) +vint8m1_t vand(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m1_m))) +vint8m1_t vand(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m2))) +vint8m2_t vand(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m2_m))) +vint8m2_t vand(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m4))) +vint8m4_t vand(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m4_m))) +vint8m4_t vand(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m8))) +vint8m8_t vand(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m8_m))) +vint8m8_t vand(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8mf2))) +vint8mf2_t vand(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8mf2_m))) +vint8mf2_t vand(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8mf4))) +vint8mf4_t vand(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8mf4_m))) +vint8mf4_t vand(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8mf8))) +vint8mf8_t vand(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8mf8_m))) +vint8mf8_t vand(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m1))) +vint16m1_t vand(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m1_m))) +vint16m1_t vand(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m2))) +vint16m2_t vand(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m2_m))) +vint16m2_t vand(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m4))) +vint16m4_t vand(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m4_m))) +vint16m4_t vand(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m8))) +vint16m8_t vand(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m8_m))) +vint16m8_t vand(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16mf2))) +vint16mf2_t vand(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16mf2_m))) +vint16mf2_t vand(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16mf4))) +vint16mf4_t vand(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16mf4_m))) +vint16mf4_t vand(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m1))) +vint32m1_t vand(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m1_m))) +vint32m1_t vand(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m2))) +vint32m2_t vand(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m2_m))) +vint32m2_t vand(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m4))) +vint32m4_t vand(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m4_m))) +vint32m4_t vand(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m8))) +vint32m8_t vand(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m8_m))) +vint32m8_t vand(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32mf2))) +vint32mf2_t vand(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32mf2_m))) +vint32mf2_t vand(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m1))) +vint64m1_t vand(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m1_m))) +vint64m1_t vand(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m2))) +vint64m2_t vand(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m2_m))) +vint64m2_t vand(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m4))) +vint64m4_t vand(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m4_m))) +vint64m4_t vand(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m8))) +vint64m8_t vand(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m8_m))) +vint64m8_t vand(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m1))) +vint8m1_t vand(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m1_m))) +vint8m1_t vand(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m2))) +vint8m2_t vand(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m2_m))) +vint8m2_t vand(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m4))) +vint8m4_t vand(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m4_m))) +vint8m4_t vand(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m8))) +vint8m8_t vand(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m8_m))) +vint8m8_t vand(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8mf2))) +vint8mf2_t vand(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8mf2_m))) +vint8mf2_t vand(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8mf4))) +vint8mf4_t vand(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8mf4_m))) +vint8mf4_t vand(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8mf8))) +vint8mf8_t vand(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8mf8_m))) +vint8mf8_t vand(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m1))) +vint16m1_t vand(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m1_m))) +vint16m1_t vand(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m2))) +vint16m2_t vand(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m2_m))) +vint16m2_t vand(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m4))) +vint16m4_t vand(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m4_m))) +vint16m4_t vand(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m8))) +vint16m8_t vand(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m8_m))) +vint16m8_t vand(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16mf2))) +vint16mf2_t vand(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16mf2_m))) +vint16mf2_t vand(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16mf4))) +vint16mf4_t vand(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16mf4_m))) +vint16mf4_t vand(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m1))) +vint32m1_t vand(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m1_m))) +vint32m1_t vand(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m2))) +vint32m2_t vand(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m2_m))) +vint32m2_t vand(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m4))) +vint32m4_t vand(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m4_m))) +vint32m4_t vand(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m8))) +vint32m8_t vand(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m8_m))) +vint32m8_t vand(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32mf2))) +vint32mf2_t vand(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32mf2_m))) +vint32mf2_t vand(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m1))) +vint64m1_t vand(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m1_m))) +vint64m1_t vand(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m2))) +vint64m2_t vand(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m2_m))) +vint64m2_t vand(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m4))) +vint64m4_t vand(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m4_m))) +vint64m4_t vand(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m8))) +vint64m8_t vand(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m8_m))) +vint64m8_t vand(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m1))) +vuint8m1_t vand(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m1_m))) +vuint8m1_t vand(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m2))) +vuint8m2_t vand(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m2_m))) +vuint8m2_t vand(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m4))) +vuint8m4_t vand(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m4_m))) +vuint8m4_t vand(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m8))) +vuint8m8_t vand(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m8_m))) +vuint8m8_t vand(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8mf2))) +vuint8mf2_t vand(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8mf2_m))) +vuint8mf2_t vand(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8mf4))) +vuint8mf4_t vand(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8mf4_m))) +vuint8mf4_t vand(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8mf8))) +vuint8mf8_t vand(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8mf8_m))) +vuint8mf8_t vand(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m1))) +vuint16m1_t vand(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m1_m))) +vuint16m1_t vand(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m2))) +vuint16m2_t vand(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m2_m))) +vuint16m2_t vand(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m4))) +vuint16m4_t vand(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m4_m))) +vuint16m4_t vand(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m8))) +vuint16m8_t vand(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m8_m))) +vuint16m8_t vand(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16mf2))) +vuint16mf2_t vand(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16mf2_m))) +vuint16mf2_t vand(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16mf4))) +vuint16mf4_t vand(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16mf4_m))) +vuint16mf4_t vand(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m1))) +vuint32m1_t vand(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m1_m))) +vuint32m1_t vand(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m2))) +vuint32m2_t vand(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m2_m))) +vuint32m2_t vand(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m4))) +vuint32m4_t vand(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m4_m))) +vuint32m4_t vand(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m8))) +vuint32m8_t vand(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m8_m))) +vuint32m8_t vand(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32mf2))) +vuint32mf2_t vand(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32mf2_m))) +vuint32mf2_t vand(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m1))) +vuint64m1_t vand(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m1_m))) +vuint64m1_t vand(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m2))) +vuint64m2_t vand(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m2_m))) +vuint64m2_t vand(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m4))) +vuint64m4_t vand(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m4_m))) +vuint64m4_t vand(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m8))) +vuint64m8_t vand(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m8_m))) +vuint64m8_t vand(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m1))) +vuint8m1_t vand(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m1_m))) +vuint8m1_t vand(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m2))) +vuint8m2_t vand(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m2_m))) +vuint8m2_t vand(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m4))) +vuint8m4_t vand(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m4_m))) +vuint8m4_t vand(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m8))) +vuint8m8_t vand(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m8_m))) +vuint8m8_t vand(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8mf2))) +vuint8mf2_t vand(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8mf2_m))) +vuint8mf2_t vand(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8mf4))) +vuint8mf4_t vand(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8mf4_m))) +vuint8mf4_t vand(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8mf8))) +vuint8mf8_t vand(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8mf8_m))) +vuint8mf8_t vand(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m1))) +vuint16m1_t vand(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m1_m))) +vuint16m1_t vand(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m2))) +vuint16m2_t vand(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m2_m))) +vuint16m2_t vand(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m4))) +vuint16m4_t vand(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m4_m))) +vuint16m4_t vand(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m8))) +vuint16m8_t vand(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m8_m))) +vuint16m8_t vand(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16mf2))) +vuint16mf2_t vand(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16mf2_m))) +vuint16mf2_t vand(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16mf4))) +vuint16mf4_t vand(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16mf4_m))) +vuint16mf4_t vand(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m1))) +vuint32m1_t vand(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m1_m))) +vuint32m1_t vand(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m2))) +vuint32m2_t vand(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m2_m))) +vuint32m2_t vand(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m4))) +vuint32m4_t vand(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m4_m))) +vuint32m4_t vand(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m8))) +vuint32m8_t vand(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m8_m))) +vuint32m8_t vand(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32mf2))) +vuint32mf2_t vand(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32mf2_m))) +vuint32mf2_t vand(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m1))) +vuint64m1_t vand(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m1_m))) +vuint64m1_t vand(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m2))) +vuint64m2_t vand(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m2_m))) +vuint64m2_t vand(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m4))) +vuint64m4_t vand(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m4_m))) +vuint64m4_t vand(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m8))) +vuint64m8_t vand(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m8_m))) +vuint64m8_t vand(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m1))) +vint8m1_t vxor(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m1_m))) +vint8m1_t vxor(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m2))) +vint8m2_t vxor(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m2_m))) +vint8m2_t vxor(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m4))) +vint8m4_t vxor(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m4_m))) +vint8m4_t vxor(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m8))) +vint8m8_t vxor(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m8_m))) +vint8m8_t vxor(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8mf2))) +vint8mf2_t vxor(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8mf2_m))) +vint8mf2_t vxor(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8mf4))) +vint8mf4_t vxor(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8mf4_m))) +vint8mf4_t vxor(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8mf8))) +vint8mf8_t vxor(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8mf8_m))) +vint8mf8_t vxor(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m1))) +vint16m1_t vxor(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m1_m))) +vint16m1_t vxor(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m2))) +vint16m2_t vxor(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m2_m))) +vint16m2_t vxor(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m4))) +vint16m4_t vxor(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m4_m))) +vint16m4_t vxor(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m8))) +vint16m8_t vxor(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m8_m))) +vint16m8_t vxor(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16mf2))) +vint16mf2_t vxor(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16mf2_m))) +vint16mf2_t vxor(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16mf4))) +vint16mf4_t vxor(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16mf4_m))) +vint16mf4_t vxor(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m1))) +vint32m1_t vxor(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m1_m))) +vint32m1_t vxor(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m2))) +vint32m2_t vxor(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m2_m))) +vint32m2_t vxor(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m4))) +vint32m4_t vxor(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m4_m))) +vint32m4_t vxor(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m8))) +vint32m8_t vxor(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m8_m))) +vint32m8_t vxor(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32mf2))) +vint32mf2_t vxor(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32mf2_m))) +vint32mf2_t vxor(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m1))) +vint64m1_t vxor(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m1_m))) +vint64m1_t vxor(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m2))) +vint64m2_t vxor(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m2_m))) +vint64m2_t vxor(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m4))) +vint64m4_t vxor(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m4_m))) +vint64m4_t vxor(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m8))) +vint64m8_t vxor(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m8_m))) +vint64m8_t vxor(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m1))) +vint8m1_t vxor(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m1_m))) +vint8m1_t vxor(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m2))) +vint8m2_t vxor(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m2_m))) +vint8m2_t vxor(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m4))) +vint8m4_t vxor(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m4_m))) +vint8m4_t vxor(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m8))) +vint8m8_t vxor(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m8_m))) +vint8m8_t vxor(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8mf2))) +vint8mf2_t vxor(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8mf2_m))) +vint8mf2_t vxor(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8mf4))) +vint8mf4_t vxor(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8mf4_m))) +vint8mf4_t vxor(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8mf8))) +vint8mf8_t vxor(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8mf8_m))) +vint8mf8_t vxor(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m1))) +vint16m1_t vxor(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m1_m))) +vint16m1_t vxor(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m2))) +vint16m2_t vxor(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m2_m))) +vint16m2_t vxor(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m4))) +vint16m4_t vxor(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m4_m))) +vint16m4_t vxor(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m8))) +vint16m8_t vxor(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m8_m))) +vint16m8_t vxor(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16mf2))) +vint16mf2_t vxor(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16mf2_m))) +vint16mf2_t vxor(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16mf4))) +vint16mf4_t vxor(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16mf4_m))) +vint16mf4_t vxor(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m1))) +vint32m1_t vxor(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m1_m))) +vint32m1_t vxor(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m2))) +vint32m2_t vxor(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m2_m))) +vint32m2_t vxor(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m4))) +vint32m4_t vxor(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m4_m))) +vint32m4_t vxor(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m8))) +vint32m8_t vxor(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m8_m))) +vint32m8_t vxor(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32mf2))) +vint32mf2_t vxor(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32mf2_m))) +vint32mf2_t vxor(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m1))) +vint64m1_t vxor(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m1_m))) +vint64m1_t vxor(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m2))) +vint64m2_t vxor(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m2_m))) +vint64m2_t vxor(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m4))) +vint64m4_t vxor(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m4_m))) +vint64m4_t vxor(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m8))) +vint64m8_t vxor(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m8_m))) +vint64m8_t vxor(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m1))) +void vsuxei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m1_m))) +void vsuxei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m2))) +void vsuxei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m2_m))) +void vsuxei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m4))) +void vsuxei16(int16_t * op0, vuint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m4_m))) +void vsuxei16(vbool4_t op0, int16_t * op1, vuint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m8))) +void vsuxei16(int16_t * op0, vuint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m8_m))) +void vsuxei16(vbool2_t op0, int16_t * op1, vuint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16mf2))) +void vsuxei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16mf2_m))) +void vsuxei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16mf4))) +void vsuxei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16mf4_m))) +void vsuxei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m1))) +vuint8m1_t vxor(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m1_m))) +vuint8m1_t vxor(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m2))) +vuint8m2_t vxor(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m2_m))) +vuint8m2_t vxor(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m4))) +vuint8m4_t vxor(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m4_m))) +vuint8m4_t vxor(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m8))) +vuint8m8_t vxor(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m8_m))) +vuint8m8_t vxor(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8mf2))) +vuint8mf2_t vxor(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8mf2_m))) +vuint8mf2_t vxor(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8mf4))) +vuint8mf4_t vxor(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8mf4_m))) +vuint8mf4_t vxor(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8mf8))) +vuint8mf8_t vxor(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8mf8_m))) +vuint8mf8_t vxor(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m1))) +vuint16m1_t vxor(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m1_m))) +vuint16m1_t vxor(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m2))) +vuint16m2_t vxor(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m2_m))) +vuint16m2_t vxor(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m4))) +vuint16m4_t vxor(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m4_m))) +vuint16m4_t vxor(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m8))) +vuint16m8_t vxor(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m8_m))) +vuint16m8_t vxor(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16mf2))) +vuint16mf2_t vxor(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16mf2_m))) +vuint16mf2_t vxor(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16mf4))) +vuint16mf4_t vxor(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16mf4_m))) +vuint16mf4_t vxor(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m1))) +vuint32m1_t vxor(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m1_m))) +vuint32m1_t vxor(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m2))) +vuint32m2_t vxor(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m2_m))) +vuint32m2_t vxor(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m4))) +vuint32m4_t vxor(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m4_m))) +vuint32m4_t vxor(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m8))) +vuint32m8_t vxor(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m8_m))) +vuint32m8_t vxor(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32mf2))) +vuint32mf2_t vxor(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32mf2_m))) +vuint32mf2_t vxor(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m1))) +vuint64m1_t vxor(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m1_m))) +vuint64m1_t vxor(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m2))) +vuint64m2_t vxor(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m2_m))) +vuint64m2_t vxor(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m4))) +vuint64m4_t vxor(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m4_m))) +vuint64m4_t vxor(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m8))) +vuint64m8_t vxor(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m8_m))) +vuint64m8_t vxor(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m1))) +vuint8m1_t vxor(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m1_m))) +vuint8m1_t vxor(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m2))) +vuint8m2_t vxor(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m2_m))) +vuint8m2_t vxor(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m4))) +vuint8m4_t vxor(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m4_m))) +vuint8m4_t vxor(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m8))) +vuint8m8_t vxor(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m8_m))) +vuint8m8_t vxor(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8mf2))) +vuint8mf2_t vxor(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8mf2_m))) +vuint8mf2_t vxor(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8mf4))) +vuint8mf4_t vxor(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8mf4_m))) +vuint8mf4_t vxor(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8mf8))) +vuint8mf8_t vxor(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8mf8_m))) +vuint8mf8_t vxor(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m1))) +vuint16m1_t vxor(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m1_m))) +vuint16m1_t vxor(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m2))) +vuint16m2_t vxor(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m2_m))) +vuint16m2_t vxor(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m4))) +vuint16m4_t vxor(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m4_m))) +vuint16m4_t vxor(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m8))) +vuint16m8_t vxor(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m8_m))) +vuint16m8_t vxor(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16mf2))) +vuint16mf2_t vxor(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16mf2_m))) +vuint16mf2_t vxor(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16mf4))) +vuint16mf4_t vxor(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16mf4_m))) +vuint16mf4_t vxor(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m1))) +vuint32m1_t vxor(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m1_m))) +vuint32m1_t vxor(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m2))) +vuint32m2_t vxor(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m2_m))) +vuint32m2_t vxor(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m4))) +vuint32m4_t vxor(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m4_m))) +vuint32m4_t vxor(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m8))) +vuint32m8_t vxor(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m8_m))) +vuint32m8_t vxor(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32mf2))) +vuint32mf2_t vxor(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32mf2_m))) +vuint32mf2_t vxor(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m1))) +vuint64m1_t vxor(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m1_m))) +vuint64m1_t vxor(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m2))) +vuint64m2_t vxor(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m2_m))) +vuint64m2_t vxor(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m4))) +vuint64m4_t vxor(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m4_m))) +vuint64m4_t vxor(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m8))) +vuint64m8_t vxor(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m8_m))) +vuint64m8_t vxor(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m1))) +vint8m1_t vor(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m1_m))) +vint8m1_t vor(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m2))) +vint8m2_t vor(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m2_m))) +vint8m2_t vor(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m4))) +vint8m4_t vor(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m4_m))) +vint8m4_t vor(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m8))) +vint8m8_t vor(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m8_m))) +vint8m8_t vor(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8mf2))) +vint8mf2_t vor(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8mf2_m))) +vint8mf2_t vor(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8mf4))) +vint8mf4_t vor(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8mf4_m))) +vint8mf4_t vor(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8mf8))) +vint8mf8_t vor(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8mf8_m))) +vint8mf8_t vor(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m1))) +vint16m1_t vor(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m1_m))) +vint16m1_t vor(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m2))) +vint16m2_t vor(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m2_m))) +vint16m2_t vor(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m4))) +vint16m4_t vor(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m4_m))) +vint16m4_t vor(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m8))) +vint16m8_t vor(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m8_m))) +vint16m8_t vor(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16mf2))) +vint16mf2_t vor(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16mf2_m))) +vint16mf2_t vor(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16mf4))) +vint16mf4_t vor(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16mf4_m))) +vint16mf4_t vor(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m1))) +vint32m1_t vor(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m1_m))) +vint32m1_t vor(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m2))) +vint32m2_t vor(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m2_m))) +vint32m2_t vor(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m4))) +vint32m4_t vor(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m4_m))) +vint32m4_t vor(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m8))) +vint32m8_t vor(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m8_m))) +vint32m8_t vor(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32mf2))) +vint32mf2_t vor(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32mf2_m))) +vint32mf2_t vor(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m1))) +vint64m1_t vor(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m1_m))) +vint64m1_t vor(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m2))) +vint64m2_t vor(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m2_m))) +vint64m2_t vor(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m4))) +vint64m4_t vor(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m4_m))) +vint64m4_t vor(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m8))) +vint64m8_t vor(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m8_m))) +vint64m8_t vor(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m1))) +vint8m1_t vor(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m1_m))) +vint8m1_t vor(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m2))) +vint8m2_t vor(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m2_m))) +vint8m2_t vor(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m4))) +vint8m4_t vor(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m4_m))) +vint8m4_t vor(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m8))) +vint8m8_t vor(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m8_m))) +vint8m8_t vor(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8mf2))) +vint8mf2_t vor(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8mf2_m))) +vint8mf2_t vor(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8mf4))) +vint8mf4_t vor(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8mf4_m))) +vint8mf4_t vor(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8mf8))) +vint8mf8_t vor(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8mf8_m))) +vint8mf8_t vor(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m1))) +vint16m1_t vor(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m1_m))) +vint16m1_t vor(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m2))) +vint16m2_t vor(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m2_m))) +vint16m2_t vor(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m4))) +vint16m4_t vor(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m4_m))) +vint16m4_t vor(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m8))) +vint16m8_t vor(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m8_m))) +vint16m8_t vor(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16mf2))) +vint16mf2_t vor(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16mf2_m))) +vint16mf2_t vor(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16mf4))) +vint16mf4_t vor(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16mf4_m))) +vint16mf4_t vor(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m1))) +vint32m1_t vor(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m1_m))) +vint32m1_t vor(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m2))) +vint32m2_t vor(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m2_m))) +vint32m2_t vor(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m4))) +vint32m4_t vor(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m4_m))) +vint32m4_t vor(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m8))) +vint32m8_t vor(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m8_m))) +vint32m8_t vor(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32mf2))) +vint32mf2_t vor(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32mf2_m))) +vint32mf2_t vor(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m1))) +vint64m1_t vor(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m1_m))) +vint64m1_t vor(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m2))) +vint64m2_t vor(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m2_m))) +vint64m2_t vor(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m4))) +vint64m4_t vor(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m4_m))) +vint64m4_t vor(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m8))) +vint64m8_t vor(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m8_m))) +vint64m8_t vor(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m1))) +vuint8m1_t vor(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m1_m))) +vuint8m1_t vor(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m2))) +vuint8m2_t vor(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m2_m))) +vuint8m2_t vor(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m4))) +vuint8m4_t vor(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m4_m))) +vuint8m4_t vor(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m8))) +vuint8m8_t vor(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m8_m))) +vuint8m8_t vor(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8mf2))) +vuint8mf2_t vor(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8mf2_m))) +vuint8mf2_t vor(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8mf4))) +vuint8mf4_t vor(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8mf4_m))) +vuint8mf4_t vor(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8mf8))) +vuint8mf8_t vor(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8mf8_m))) +vuint8mf8_t vor(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m1))) +vuint16m1_t vor(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m1_m))) +vuint16m1_t vor(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m2))) +vuint16m2_t vor(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m2_m))) +vuint16m2_t vor(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m4))) +vuint16m4_t vor(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m4_m))) +vuint16m4_t vor(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m8))) +vuint16m8_t vor(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m8_m))) +vuint16m8_t vor(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16mf2))) +vuint16mf2_t vor(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16mf2_m))) +vuint16mf2_t vor(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16mf4))) +vuint16mf4_t vor(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16mf4_m))) +vuint16mf4_t vor(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m1))) +vuint32m1_t vor(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m1_m))) +vuint32m1_t vor(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m2))) +vuint32m2_t vor(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m2_m))) +vuint32m2_t vor(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m4))) +vuint32m4_t vor(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m4_m))) +vuint32m4_t vor(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m8))) +vuint32m8_t vor(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m8_m))) +vuint32m8_t vor(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32mf2))) +vuint32mf2_t vor(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32mf2_m))) +vuint32mf2_t vor(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m1))) +vuint64m1_t vor(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m1_m))) +vuint64m1_t vor(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m2))) +vuint64m2_t vor(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m2_m))) +vuint64m2_t vor(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m4))) +vuint64m4_t vor(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m4_m))) +vuint64m4_t vor(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m8))) +vuint64m8_t vor(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m8_m))) +vuint64m8_t vor(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m1))) +vuint8m1_t vor(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m1_m))) +vuint8m1_t vor(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m2))) +vuint8m2_t vor(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m2_m))) +vuint8m2_t vor(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m4))) +vuint8m4_t vor(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m4_m))) +vuint8m4_t vor(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m8))) +vuint8m8_t vor(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m8_m))) +vuint8m8_t vor(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8mf2))) +vuint8mf2_t vor(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8mf2_m))) +vuint8mf2_t vor(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8mf4))) +vuint8mf4_t vor(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8mf4_m))) +vuint8mf4_t vor(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8mf8))) +vuint8mf8_t vor(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8mf8_m))) +vuint8mf8_t vor(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m1))) +vuint16m1_t vor(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m1_m))) +vuint16m1_t vor(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m2))) +vuint16m2_t vor(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m2_m))) +vuint16m2_t vor(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m4))) +vuint16m4_t vor(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m4_m))) +vuint16m4_t vor(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m8))) +vuint16m8_t vor(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m8_m))) +vuint16m8_t vor(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16mf2))) +vuint16mf2_t vor(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16mf2_m))) +vuint16mf2_t vor(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16mf4))) +vuint16mf4_t vor(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16mf4_m))) +vuint16mf4_t vor(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m1))) +vuint32m1_t vor(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m1_m))) +vuint32m1_t vor(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m2))) +vuint32m2_t vor(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m2_m))) +vuint32m2_t vor(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m4))) +vuint32m4_t vor(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m4_m))) +vuint32m4_t vor(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m8))) +vuint32m8_t vor(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m8_m))) +vuint32m8_t vor(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32mf2))) +vuint32mf2_t vor(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32mf2_m))) +vuint32mf2_t vor(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m1))) +vuint64m1_t vor(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m1_m))) +vuint64m1_t vor(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m2))) +vuint64m2_t vor(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m2_m))) +vuint64m2_t vor(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m4))) +vuint64m4_t vor(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m4_m))) +vuint64m4_t vor(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m8))) +vuint64m8_t vor(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m8_m))) +vuint64m8_t vor(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m1))) +vint8m1_t vsll(vint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m1_m))) +vint8m1_t vsll(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m2))) +vint8m2_t vsll(vint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m2_m))) +vint8m2_t vsll(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m4))) +vint8m4_t vsll(vint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m4_m))) +vint8m4_t vsll(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m8))) +vint8m8_t vsll(vint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m8_m))) +vint8m8_t vsll(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8mf2))) +vint8mf2_t vsll(vint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8mf2_m))) +vint8mf2_t vsll(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8mf4))) +vint8mf4_t vsll(vint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8mf4_m))) +vint8mf4_t vsll(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8mf8))) +vint8mf8_t vsll(vint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8mf8_m))) +vint8mf8_t vsll(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m1))) +vint16m1_t vsll(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m1_m))) +vint16m1_t vsll(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m2))) +vint16m2_t vsll(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m2_m))) +vint16m2_t vsll(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m4))) +vint16m4_t vsll(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m4_m))) +vint16m4_t vsll(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m8))) +vint16m8_t vsll(vint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m8_m))) +vint16m8_t vsll(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16mf2))) +vint16mf2_t vsll(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16mf2_m))) +vint16mf2_t vsll(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16mf4))) +vint16mf4_t vsll(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16mf4_m))) +vint16mf4_t vsll(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m1))) +vint32m1_t vsll(vint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m1_m))) +vint32m1_t vsll(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m2))) +vint32m2_t vsll(vint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m2_m))) +vint32m2_t vsll(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m4))) +vint32m4_t vsll(vint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m4_m))) +vint32m4_t vsll(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m8))) +vint32m8_t vsll(vint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m8_m))) +vint32m8_t vsll(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32mf2))) +vint32mf2_t vsll(vint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32mf2_m))) +vint32mf2_t vsll(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m1))) +vint64m1_t vsll(vint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m1_m))) +vint64m1_t vsll(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m2))) +vint64m2_t vsll(vint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m2_m))) +vint64m2_t vsll(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m4))) +vint64m4_t vsll(vint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m4_m))) +vint64m4_t vsll(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m8))) +vint64m8_t vsll(vint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m8_m))) +vint64m8_t vsll(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m1))) +vint8m1_t vsll(vint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m1_m))) +vint8m1_t vsll(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m2))) +vint8m2_t vsll(vint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m2_m))) +vint8m2_t vsll(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m4))) +vint8m4_t vsll(vint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m4_m))) +vint8m4_t vsll(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m8))) +vint8m8_t vsll(vint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m8_m))) +vint8m8_t vsll(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8mf2))) +vint8mf2_t vsll(vint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8mf2_m))) +vint8mf2_t vsll(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8mf4))) +vint8mf4_t vsll(vint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8mf4_m))) +vint8mf4_t vsll(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8mf8))) +vint8mf8_t vsll(vint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8mf8_m))) +vint8mf8_t vsll(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m1))) +vint16m1_t vsll(vint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m1_m))) +vint16m1_t vsll(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m2))) +vint16m2_t vsll(vint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m2_m))) +vint16m2_t vsll(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m4))) +vint16m4_t vsll(vint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m4_m))) +vint16m4_t vsll(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m8))) +vint16m8_t vsll(vint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m8_m))) +vint16m8_t vsll(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16mf2))) +vint16mf2_t vsll(vint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16mf2_m))) +vint16mf2_t vsll(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16mf4))) +vint16mf4_t vsll(vint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16mf4_m))) +vint16mf4_t vsll(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m1))) +vint32m1_t vsll(vint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m1_m))) +vint32m1_t vsll(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m2))) +vint32m2_t vsll(vint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m2_m))) +vint32m2_t vsll(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m4))) +vint32m4_t vsll(vint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m4_m))) +vint32m4_t vsll(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m8))) +vint32m8_t vsll(vint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m8_m))) +vint32m8_t vsll(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32mf2))) +vint32mf2_t vsll(vint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32mf2_m))) +vint32mf2_t vsll(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m1))) +vint64m1_t vsll(vint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m1_m))) +vint64m1_t vsll(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m2))) +vint64m2_t vsll(vint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m2_m))) +vint64m2_t vsll(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m4))) +vint64m4_t vsll(vint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m4_m))) +vint64m4_t vsll(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m8))) +vint64m8_t vsll(vint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m8_m))) +vint64m8_t vsll(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m1))) +vuint8m1_t vsll(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m1_m))) +vuint8m1_t vsll(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m2))) +vuint8m2_t vsll(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m2_m))) +vuint8m2_t vsll(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m4))) +vuint8m4_t vsll(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m4_m))) +vuint8m4_t vsll(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m8))) +vuint8m8_t vsll(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m8_m))) +vuint8m8_t vsll(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8mf2))) +vuint8mf2_t vsll(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8mf2_m))) +vuint8mf2_t vsll(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8mf4))) +vuint8mf4_t vsll(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8mf4_m))) +vuint8mf4_t vsll(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8mf8))) +vuint8mf8_t vsll(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8mf8_m))) +vuint8mf8_t vsll(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m1))) +vuint16m1_t vsll(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m1_m))) +vuint16m1_t vsll(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m2))) +vuint16m2_t vsll(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m2_m))) +vuint16m2_t vsll(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m4))) +vuint16m4_t vsll(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m4_m))) +vuint16m4_t vsll(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m8))) +vuint16m8_t vsll(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m8_m))) +vuint16m8_t vsll(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16mf2))) +vuint16mf2_t vsll(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16mf2_m))) +vuint16mf2_t vsll(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16mf4))) +vuint16mf4_t vsll(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16mf4_m))) +vuint16mf4_t vsll(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m1))) +vuint32m1_t vsll(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m1_m))) +vuint32m1_t vsll(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m2))) +vuint32m2_t vsll(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m2_m))) +vuint32m2_t vsll(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m4))) +vuint32m4_t vsll(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m4_m))) +vuint32m4_t vsll(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m8))) +vuint32m8_t vsll(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m8_m))) +vuint32m8_t vsll(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32mf2))) +vuint32mf2_t vsll(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32mf2_m))) +vuint32mf2_t vsll(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m1))) +vuint64m1_t vsll(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m1_m))) +vuint64m1_t vsll(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m2))) +vuint64m2_t vsll(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m2_m))) +vuint64m2_t vsll(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m4))) +vuint64m4_t vsll(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m4_m))) +vuint64m4_t vsll(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m8))) +vuint64m8_t vsll(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m8_m))) +vuint64m8_t vsll(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m1))) +vuint8m1_t vsll(vuint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m1_m))) +vuint8m1_t vsll(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m2))) +vuint8m2_t vsll(vuint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m2_m))) +vuint8m2_t vsll(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m4))) +vuint8m4_t vsll(vuint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m4_m))) +vuint8m4_t vsll(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m8))) +vuint8m8_t vsll(vuint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m8_m))) +vuint8m8_t vsll(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8mf2))) +vuint8mf2_t vsll(vuint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8mf2_m))) +vuint8mf2_t vsll(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8mf4))) +vuint8mf4_t vsll(vuint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8mf4_m))) +vuint8mf4_t vsll(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8mf8))) +vuint8mf8_t vsll(vuint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8mf8_m))) +vuint8mf8_t vsll(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m1))) +vuint16m1_t vsll(vuint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m1_m))) +vuint16m1_t vsll(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m2))) +vuint16m2_t vsll(vuint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m2_m))) +vuint16m2_t vsll(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m4))) +vuint16m4_t vsll(vuint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m4_m))) +vuint16m4_t vsll(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m8))) +vuint16m8_t vsll(vuint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m8_m))) +vuint16m8_t vsll(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16mf2))) +vuint16mf2_t vsll(vuint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16mf2_m))) +vuint16mf2_t vsll(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16mf4))) +vuint16mf4_t vsll(vuint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16mf4_m))) +vuint16mf4_t vsll(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m1))) +vuint32m1_t vsll(vuint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m1_m))) +vuint32m1_t vsll(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m2))) +vuint32m2_t vsll(vuint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m2_m))) +vuint32m2_t vsll(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m4))) +vuint32m4_t vsll(vuint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m4_m))) +vuint32m4_t vsll(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m8))) +vuint32m8_t vsll(vuint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m8_m))) +vuint32m8_t vsll(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32mf2))) +vuint32mf2_t vsll(vuint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32mf2_m))) +vuint32mf2_t vsll(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m1))) +vuint64m1_t vsll(vuint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m1_m))) +vuint64m1_t vsll(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m2))) +vuint64m2_t vsll(vuint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m2_m))) +vuint64m2_t vsll(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m4))) +vuint64m4_t vsll(vuint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m4_m))) +vuint64m4_t vsll(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m8))) +vuint64m8_t vsll(vuint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m8_m))) +vuint64m8_t vsll(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m1))) +void vsuxei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m1_m))) +void vsuxei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m2))) +void vsuxei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m2_m))) +void vsuxei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m4))) +void vsuxei16(uint16_t * op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m4_m))) +void vsuxei16(vbool4_t op0, uint16_t * op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m8))) +void vsuxei16(uint16_t * op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m8_m))) +void vsuxei16(vbool2_t op0, uint16_t * op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16mf2))) +void vsuxei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16mf2_m))) +void vsuxei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16mf4))) +void vsuxei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16mf4_m))) +void vsuxei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m1))) +vuint8m1_t vsrl(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m1_m))) +vuint8m1_t vsrl(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m2))) +vuint8m2_t vsrl(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m2_m))) +vuint8m2_t vsrl(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m4))) +vuint8m4_t vsrl(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m4_m))) +vuint8m4_t vsrl(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m8))) +vuint8m8_t vsrl(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m8_m))) +vuint8m8_t vsrl(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8mf2))) +vuint8mf2_t vsrl(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8mf2_m))) +vuint8mf2_t vsrl(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8mf4))) +vuint8mf4_t vsrl(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8mf4_m))) +vuint8mf4_t vsrl(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8mf8))) +vuint8mf8_t vsrl(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8mf8_m))) +vuint8mf8_t vsrl(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m1))) +vuint16m1_t vsrl(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m1_m))) +vuint16m1_t vsrl(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m2))) +vuint16m2_t vsrl(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m2_m))) +vuint16m2_t vsrl(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m4))) +vuint16m4_t vsrl(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m4_m))) +vuint16m4_t vsrl(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m8))) +vuint16m8_t vsrl(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m8_m))) +vuint16m8_t vsrl(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16mf2))) +vuint16mf2_t vsrl(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16mf2_m))) +vuint16mf2_t vsrl(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16mf4))) +vuint16mf4_t vsrl(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16mf4_m))) +vuint16mf4_t vsrl(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m1))) +vuint32m1_t vsrl(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m1_m))) +vuint32m1_t vsrl(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m2))) +vuint32m2_t vsrl(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m2_m))) +vuint32m2_t vsrl(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m4))) +vuint32m4_t vsrl(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m4_m))) +vuint32m4_t vsrl(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m8))) +vuint32m8_t vsrl(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m8_m))) +vuint32m8_t vsrl(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32mf2))) +vuint32mf2_t vsrl(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32mf2_m))) +vuint32mf2_t vsrl(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m1))) +vuint64m1_t vsrl(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m1_m))) +vuint64m1_t vsrl(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m2))) +vuint64m2_t vsrl(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m2_m))) +vuint64m2_t vsrl(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m4))) +vuint64m4_t vsrl(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m4_m))) +vuint64m4_t vsrl(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m8))) +vuint64m8_t vsrl(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m8_m))) +vuint64m8_t vsrl(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m1))) +vuint8m1_t vsrl(vuint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m1_m))) +vuint8m1_t vsrl(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m2))) +vuint8m2_t vsrl(vuint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m2_m))) +vuint8m2_t vsrl(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m4))) +vuint8m4_t vsrl(vuint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m4_m))) +vuint8m4_t vsrl(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m8))) +vuint8m8_t vsrl(vuint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m8_m))) +vuint8m8_t vsrl(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8mf2))) +vuint8mf2_t vsrl(vuint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8mf2_m))) +vuint8mf2_t vsrl(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8mf4))) +vuint8mf4_t vsrl(vuint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8mf4_m))) +vuint8mf4_t vsrl(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8mf8))) +vuint8mf8_t vsrl(vuint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8mf8_m))) +vuint8mf8_t vsrl(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m1))) +vuint16m1_t vsrl(vuint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m1_m))) +vuint16m1_t vsrl(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m2))) +vuint16m2_t vsrl(vuint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m2_m))) +vuint16m2_t vsrl(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m4))) +vuint16m4_t vsrl(vuint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m4_m))) +vuint16m4_t vsrl(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m8))) +vuint16m8_t vsrl(vuint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m8_m))) +vuint16m8_t vsrl(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16mf2))) +vuint16mf2_t vsrl(vuint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16mf2_m))) +vuint16mf2_t vsrl(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16mf4))) +vuint16mf4_t vsrl(vuint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16mf4_m))) +vuint16mf4_t vsrl(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m1))) +vuint32m1_t vsrl(vuint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m1_m))) +vuint32m1_t vsrl(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m2))) +vuint32m2_t vsrl(vuint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m2_m))) +vuint32m2_t vsrl(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m4))) +vuint32m4_t vsrl(vuint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m4_m))) +vuint32m4_t vsrl(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m8))) +vuint32m8_t vsrl(vuint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m8_m))) +vuint32m8_t vsrl(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32mf2))) +vuint32mf2_t vsrl(vuint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32mf2_m))) +vuint32mf2_t vsrl(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m1))) +vuint64m1_t vsrl(vuint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m1_m))) +vuint64m1_t vsrl(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m2))) +vuint64m2_t vsrl(vuint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m2_m))) +vuint64m2_t vsrl(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m4))) +vuint64m4_t vsrl(vuint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m4_m))) +vuint64m4_t vsrl(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m8))) +vuint64m8_t vsrl(vuint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m8_m))) +vuint64m8_t vsrl(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m1))) +vint8m1_t vsra(vint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m1_m))) +vint8m1_t vsra(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m2))) +vint8m2_t vsra(vint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m2_m))) +vint8m2_t vsra(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m4))) +vint8m4_t vsra(vint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m4_m))) +vint8m4_t vsra(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m8))) +vint8m8_t vsra(vint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m8_m))) +vint8m8_t vsra(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8mf2))) +vint8mf2_t vsra(vint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8mf2_m))) +vint8mf2_t vsra(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8mf4))) +vint8mf4_t vsra(vint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8mf4_m))) +vint8mf4_t vsra(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8mf8))) +vint8mf8_t vsra(vint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8mf8_m))) +vint8mf8_t vsra(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m1))) +vint16m1_t vsra(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m1_m))) +vint16m1_t vsra(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m2))) +vint16m2_t vsra(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m2_m))) +vint16m2_t vsra(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m4))) +vint16m4_t vsra(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m4_m))) +vint16m4_t vsra(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m8))) +vint16m8_t vsra(vint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m8_m))) +vint16m8_t vsra(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16mf2))) +vint16mf2_t vsra(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16mf2_m))) +vint16mf2_t vsra(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16mf4))) +vint16mf4_t vsra(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16mf4_m))) +vint16mf4_t vsra(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m1))) +vint32m1_t vsra(vint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m1_m))) +vint32m1_t vsra(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m2))) +vint32m2_t vsra(vint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m2_m))) +vint32m2_t vsra(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m4))) +vint32m4_t vsra(vint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m4_m))) +vint32m4_t vsra(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m8))) +vint32m8_t vsra(vint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m8_m))) +vint32m8_t vsra(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32mf2))) +vint32mf2_t vsra(vint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32mf2_m))) +vint32mf2_t vsra(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m1))) +vint64m1_t vsra(vint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m1_m))) +vint64m1_t vsra(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m2))) +vint64m2_t vsra(vint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m2_m))) +vint64m2_t vsra(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m4))) +vint64m4_t vsra(vint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m4_m))) +vint64m4_t vsra(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m8))) +vint64m8_t vsra(vint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m8_m))) +vint64m8_t vsra(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m1))) +vint8m1_t vsra(vint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m1_m))) +vint8m1_t vsra(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m2))) +vint8m2_t vsra(vint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m2_m))) +vint8m2_t vsra(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m4))) +vint8m4_t vsra(vint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m4_m))) +vint8m4_t vsra(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m8))) +vint8m8_t vsra(vint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m8_m))) +vint8m8_t vsra(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8mf2))) +vint8mf2_t vsra(vint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8mf2_m))) +vint8mf2_t vsra(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8mf4))) +vint8mf4_t vsra(vint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8mf4_m))) +vint8mf4_t vsra(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8mf8))) +vint8mf8_t vsra(vint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8mf8_m))) +vint8mf8_t vsra(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m1))) +vint16m1_t vsra(vint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m1_m))) +vint16m1_t vsra(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m2))) +vint16m2_t vsra(vint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m2_m))) +vint16m2_t vsra(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m4))) +vint16m4_t vsra(vint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m4_m))) +vint16m4_t vsra(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m8))) +vint16m8_t vsra(vint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m8_m))) +vint16m8_t vsra(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16mf2))) +vint16mf2_t vsra(vint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16mf2_m))) +vint16mf2_t vsra(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16mf4))) +vint16mf4_t vsra(vint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16mf4_m))) +vint16mf4_t vsra(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m1))) +vint32m1_t vsra(vint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m1_m))) +vint32m1_t vsra(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m2))) +vint32m2_t vsra(vint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m2_m))) +vint32m2_t vsra(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m4))) +vint32m4_t vsra(vint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m4_m))) +vint32m4_t vsra(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m8))) +vint32m8_t vsra(vint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m8_m))) +vint32m8_t vsra(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32mf2))) +vint32mf2_t vsra(vint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32mf2_m))) +vint32mf2_t vsra(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m1))) +vint64m1_t vsra(vint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m1_m))) +vint64m1_t vsra(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m2))) +vint64m2_t vsra(vint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m2_m))) +vint64m2_t vsra(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m4))) +vint64m4_t vsra(vint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m4_m))) +vint64m4_t vsra(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m8))) +vint64m8_t vsra(vint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m8_m))) +vint64m8_t vsra(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8m1))) +vuint8m1_t vnsrl(vuint16m2_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8m1_m))) +vuint8m1_t vnsrl(vbool8_t op0, vuint8m1_t op1, vuint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8m2))) +vuint8m2_t vnsrl(vuint16m4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8m2_m))) +vuint8m2_t vnsrl(vbool4_t op0, vuint8m2_t op1, vuint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8m4))) +vuint8m4_t vnsrl(vuint16m8_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8m4_m))) +vuint8m4_t vnsrl(vbool2_t op0, vuint8m4_t op1, vuint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8mf2))) +vuint8mf2_t vnsrl(vuint16m1_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8mf2_m))) +vuint8mf2_t vnsrl(vbool16_t op0, vuint8mf2_t op1, vuint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8mf4))) +vuint8mf4_t vnsrl(vuint16mf2_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8mf4_m))) +vuint8mf4_t vnsrl(vbool32_t op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8mf8))) +vuint8mf8_t vnsrl(vuint16mf4_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8mf8_m))) +vuint8mf8_t vnsrl(vbool64_t op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16m1))) +vuint16m1_t vnsrl(vuint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16m1_m))) +vuint16m1_t vnsrl(vbool16_t op0, vuint16m1_t op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16m2))) +vuint16m2_t vnsrl(vuint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16m2_m))) +vuint16m2_t vnsrl(vbool8_t op0, vuint16m2_t op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16m4))) +vuint16m4_t vnsrl(vuint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16m4_m))) +vuint16m4_t vnsrl(vbool4_t op0, vuint16m4_t op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16mf2))) +vuint16mf2_t vnsrl(vuint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16mf2_m))) +vuint16mf2_t vnsrl(vbool32_t op0, vuint16mf2_t op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16mf4))) +vuint16mf4_t vnsrl(vuint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16mf4_m))) +vuint16mf4_t vnsrl(vbool64_t op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32m1))) +vuint32m1_t vnsrl(vuint64m2_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32m1_m))) +vuint32m1_t vnsrl(vbool32_t op0, vuint32m1_t op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32m2))) +vuint32m2_t vnsrl(vuint64m4_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32m2_m))) +vuint32m2_t vnsrl(vbool16_t op0, vuint32m2_t op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32m4))) +vuint32m4_t vnsrl(vuint64m8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32m4_m))) +vuint32m4_t vnsrl(vbool8_t op0, vuint32m4_t op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32mf2))) +vuint32mf2_t vnsrl(vuint64m1_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32mf2_m))) +vuint32mf2_t vnsrl(vbool64_t op0, vuint32mf2_t op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8m1))) +vuint8m1_t vnsrl(vuint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8m1_m))) +vuint8m1_t vnsrl(vbool8_t op0, vuint8m1_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8m2))) +vuint8m2_t vnsrl(vuint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8m2_m))) +vuint8m2_t vnsrl(vbool4_t op0, vuint8m2_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8m4))) +vuint8m4_t vnsrl(vuint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8m4_m))) +vuint8m4_t vnsrl(vbool2_t op0, vuint8m4_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8mf2))) +vuint8mf2_t vnsrl(vuint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8mf2_m))) +vuint8mf2_t vnsrl(vbool16_t op0, vuint8mf2_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8mf4))) +vuint8mf4_t vnsrl(vuint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8mf4_m))) +vuint8mf4_t vnsrl(vbool32_t op0, vuint8mf4_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8mf8))) +vuint8mf8_t vnsrl(vuint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8mf8_m))) +vuint8mf8_t vnsrl(vbool64_t op0, vuint8mf8_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16m1))) +vuint16m1_t vnsrl(vuint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16m1_m))) +vuint16m1_t vnsrl(vbool16_t op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16m2))) +vuint16m2_t vnsrl(vuint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16m2_m))) +vuint16m2_t vnsrl(vbool8_t op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16m4))) +vuint16m4_t vnsrl(vuint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16m4_m))) +vuint16m4_t vnsrl(vbool4_t op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16mf2))) +vuint16mf2_t vnsrl(vuint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16mf2_m))) +vuint16mf2_t vnsrl(vbool32_t op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16mf4))) +vuint16mf4_t vnsrl(vuint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16mf4_m))) +vuint16mf4_t vnsrl(vbool64_t op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32m1))) +vuint32m1_t vnsrl(vuint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32m1_m))) +vuint32m1_t vnsrl(vbool32_t op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32m2))) +vuint32m2_t vnsrl(vuint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32m2_m))) +vuint32m2_t vnsrl(vbool16_t op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32m4))) +vuint32m4_t vnsrl(vuint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32m4_m))) +vuint32m4_t vnsrl(vbool8_t op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32mf2))) +vuint32mf2_t vnsrl(vuint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32mf2_m))) +vuint32mf2_t vnsrl(vbool64_t op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8m1))) +vint8m1_t vnsra(vint16m2_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8m1_m))) +vint8m1_t vnsra(vbool8_t op0, vint8m1_t op1, vint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8m2))) +vint8m2_t vnsra(vint16m4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8m2_m))) +vint8m2_t vnsra(vbool4_t op0, vint8m2_t op1, vint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8m4))) +vint8m4_t vnsra(vint16m8_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8m4_m))) +vint8m4_t vnsra(vbool2_t op0, vint8m4_t op1, vint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8mf2))) +vint8mf2_t vnsra(vint16m1_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8mf2_m))) +vint8mf2_t vnsra(vbool16_t op0, vint8mf2_t op1, vint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8mf4))) +vint8mf4_t vnsra(vint16mf2_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8mf4_m))) +vint8mf4_t vnsra(vbool32_t op0, vint8mf4_t op1, vint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8mf8))) +vint8mf8_t vnsra(vint16mf4_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8mf8_m))) +vint8mf8_t vnsra(vbool64_t op0, vint8mf8_t op1, vint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16m1))) +vint16m1_t vnsra(vint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16m1_m))) +vint16m1_t vnsra(vbool16_t op0, vint16m1_t op1, vint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16m2))) +vint16m2_t vnsra(vint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16m2_m))) +vint16m2_t vnsra(vbool8_t op0, vint16m2_t op1, vint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16m4))) +vint16m4_t vnsra(vint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16m4_m))) +vint16m4_t vnsra(vbool4_t op0, vint16m4_t op1, vint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16mf2))) +vint16mf2_t vnsra(vint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16mf2_m))) +vint16mf2_t vnsra(vbool32_t op0, vint16mf2_t op1, vint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16mf4))) +vint16mf4_t vnsra(vint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16mf4_m))) +vint16mf4_t vnsra(vbool64_t op0, vint16mf4_t op1, vint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32m1))) +vint32m1_t vnsra(vint64m2_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32m1_m))) +vint32m1_t vnsra(vbool32_t op0, vint32m1_t op1, vint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32m2))) +vint32m2_t vnsra(vint64m4_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32m2_m))) +vint32m2_t vnsra(vbool16_t op0, vint32m2_t op1, vint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32m4))) +vint32m4_t vnsra(vint64m8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32m4_m))) +vint32m4_t vnsra(vbool8_t op0, vint32m4_t op1, vint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32mf2))) +vint32mf2_t vnsra(vint64m1_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32mf2_m))) +vint32mf2_t vnsra(vbool64_t op0, vint32mf2_t op1, vint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8m1))) +vint8m1_t vnsra(vint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8m1_m))) +vint8m1_t vnsra(vbool8_t op0, vint8m1_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8m2))) +vint8m2_t vnsra(vint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8m2_m))) +vint8m2_t vnsra(vbool4_t op0, vint8m2_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8m4))) +vint8m4_t vnsra(vint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8m4_m))) +vint8m4_t vnsra(vbool2_t op0, vint8m4_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8mf2))) +vint8mf2_t vnsra(vint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8mf2_m))) +vint8mf2_t vnsra(vbool16_t op0, vint8mf2_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8mf4))) +vint8mf4_t vnsra(vint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8mf4_m))) +vint8mf4_t vnsra(vbool32_t op0, vint8mf4_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8mf8))) +vint8mf8_t vnsra(vint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8mf8_m))) +vint8mf8_t vnsra(vbool64_t op0, vint8mf8_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16m1))) +vint16m1_t vnsra(vint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16m1_m))) +vint16m1_t vnsra(vbool16_t op0, vint16m1_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16m2))) +vint16m2_t vnsra(vint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16m2_m))) +vint16m2_t vnsra(vbool8_t op0, vint16m2_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16m4))) +vint16m4_t vnsra(vint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16m4_m))) +vint16m4_t vnsra(vbool4_t op0, vint16m4_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16mf2))) +vint16mf2_t vnsra(vint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16mf2_m))) +vint16mf2_t vnsra(vbool32_t op0, vint16mf2_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16mf4))) +vint16mf4_t vnsra(vint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16mf4_m))) +vint16mf4_t vnsra(vbool64_t op0, vint16mf4_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32m1))) +vint32m1_t vnsra(vint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32m1_m))) +vint32m1_t vnsra(vbool32_t op0, vint32m1_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32m2))) +vint32m2_t vnsra(vint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32m2_m))) +vint32m2_t vnsra(vbool16_t op0, vint32m2_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32m4))) +vint32m4_t vnsra(vint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32m4_m))) +vint32m4_t vnsra(vbool8_t op0, vint32m4_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32mf2))) +vint32mf2_t vnsra(vint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32mf2_m))) +vint32mf2_t vnsra(vbool64_t op0, vint32mf2_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8m1))) +vuint8m1_t vncvt_x(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8m1_m))) +vuint8m1_t vncvt_x(vbool8_t op0, vuint8m1_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8m2))) +vuint8m2_t vncvt_x(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8m2_m))) +vuint8m2_t vncvt_x(vbool4_t op0, vuint8m2_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8m4))) +vuint8m4_t vncvt_x(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8m4_m))) +vuint8m4_t vncvt_x(vbool2_t op0, vuint8m4_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8mf2))) +vuint8mf2_t vncvt_x(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8mf2_m))) +vuint8mf2_t vncvt_x(vbool16_t op0, vuint8mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8mf4))) +vuint8mf4_t vncvt_x(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8mf4_m))) +vuint8mf4_t vncvt_x(vbool32_t op0, vuint8mf4_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8mf8))) +vuint8mf8_t vncvt_x(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8mf8_m))) +vuint8mf8_t vncvt_x(vbool64_t op0, vuint8mf8_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16m1))) +vuint16m1_t vncvt_x(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16m1_m))) +vuint16m1_t vncvt_x(vbool16_t op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16m2))) +vuint16m2_t vncvt_x(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16m2_m))) +vuint16m2_t vncvt_x(vbool8_t op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16m4))) +vuint16m4_t vncvt_x(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16m4_m))) +vuint16m4_t vncvt_x(vbool4_t op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16mf2))) +vuint16mf2_t vncvt_x(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16mf2_m))) +vuint16mf2_t vncvt_x(vbool32_t op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16mf4))) +vuint16mf4_t vncvt_x(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16mf4_m))) +vuint16mf4_t vncvt_x(vbool64_t op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32m1))) +vuint32m1_t vncvt_x(vuint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32m1_m))) +vuint32m1_t vncvt_x(vbool32_t op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32m2))) +vuint32m2_t vncvt_x(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32m2_m))) +vuint32m2_t vncvt_x(vbool16_t op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32m4))) +vuint32m4_t vncvt_x(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32m4_m))) +vuint32m4_t vncvt_x(vbool8_t op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32mf2))) +vuint32mf2_t vncvt_x(vuint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32mf2_m))) +vuint32mf2_t vncvt_x(vbool64_t op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m1_b8))) +vbool8_t vmseq(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m1_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m2_b4))) +vbool4_t vmseq(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m2_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m4_b2))) +vbool2_t vmseq(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m4_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m8_b1))) +vbool1_t vmseq(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m8_b1_m))) +vbool1_t vmseq(vbool1_t op0, vbool1_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8mf2_b16))) +vbool16_t vmseq(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8mf2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8mf4_b32))) +vbool32_t vmseq(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8mf4_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8mf8_b64))) +vbool64_t vmseq(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8mf8_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m1_b16))) +vbool16_t vmseq(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m1_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m2_b8))) +vbool8_t vmseq(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m2_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m4_b4))) +vbool4_t vmseq(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m4_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m8_b2))) +vbool2_t vmseq(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m8_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16mf2_b32))) +vbool32_t vmseq(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16mf2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16mf4_b64))) +vbool64_t vmseq(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16mf4_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m1_b32))) +vbool32_t vmseq(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m1_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m2_b16))) +vbool16_t vmseq(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m4_b8))) +vbool8_t vmseq(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m4_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m8_b4))) +vbool4_t vmseq(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m8_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32mf2_b64))) +vbool64_t vmseq(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32mf2_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m1_b64))) +vbool64_t vmseq(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m1_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m2_b32))) +vbool32_t vmseq(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m4_b16))) +vbool16_t vmseq(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m4_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m8_b8))) +vbool8_t vmseq(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m8_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m1))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m1_m))) +void vsse8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m2))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m2_m))) +void vsse8(vbool4_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m4))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m4_m))) +void vsse8(vbool2_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m8))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m8_m))) +void vsse8(vbool1_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8mf2))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8mf2_m))) +void vsse8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8mf4))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8mf4_m))) +void vsse8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8mf8))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8mf8_m))) +void vsse8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16m1))) +void vsuxei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16m1_m))) +void vsuxei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16m2))) +void vsuxei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16m2_m))) +void vsuxei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16m4))) +void vsuxei32(int16_t * op0, vuint32m8_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16m4_m))) +void vsuxei32(vbool4_t op0, int16_t * op1, vuint32m8_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16mf2))) +void vsuxei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16mf2_m))) +void vsuxei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16mf4))) +void vsuxei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16mf4_m))) +void vsuxei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m1_b8))) +vbool8_t vmseq(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m1_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m2_b4))) +vbool4_t vmseq(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m2_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m4_b2))) +vbool2_t vmseq(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m4_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m8_b1))) +vbool1_t vmseq(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m8_b1_m))) +vbool1_t vmseq(vbool1_t op0, vbool1_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8mf2_b16))) +vbool16_t vmseq(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8mf2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8mf4_b32))) +vbool32_t vmseq(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8mf4_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8mf8_b64))) +vbool64_t vmseq(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8mf8_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m1_b16))) +vbool16_t vmseq(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m1_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m2_b8))) +vbool8_t vmseq(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m2_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m4_b4))) +vbool4_t vmseq(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m4_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m8_b2))) +vbool2_t vmseq(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m8_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16mf2_b32))) +vbool32_t vmseq(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16mf2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16mf4_b64))) +vbool64_t vmseq(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16mf4_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m1_b32))) +vbool32_t vmseq(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m1_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m2_b16))) +vbool16_t vmseq(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m4_b8))) +vbool8_t vmseq(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m4_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m8_b4))) +vbool4_t vmseq(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m8_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32mf2_b64))) +vbool64_t vmseq(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32mf2_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m1_b64))) +vbool64_t vmseq(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m1_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m2_b32))) +vbool32_t vmseq(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m4_b16))) +vbool16_t vmseq(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m4_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m8_b8))) +vbool8_t vmseq(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m8_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m1_b8))) +vbool8_t vmseq(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m1_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m2_b4))) +vbool4_t vmseq(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m2_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m4_b2))) +vbool2_t vmseq(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m4_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m8_b1))) +vbool1_t vmseq(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m8_b1_m))) +vbool1_t vmseq(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8mf2_b16))) +vbool16_t vmseq(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8mf2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8mf4_b32))) +vbool32_t vmseq(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8mf4_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8mf8_b64))) +vbool64_t vmseq(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8mf8_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m1_b16))) +vbool16_t vmseq(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m1_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m2_b8))) +vbool8_t vmseq(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m2_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m4_b4))) +vbool4_t vmseq(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m4_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m8_b2))) +vbool2_t vmseq(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m8_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16mf2_b32))) +vbool32_t vmseq(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16mf2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16mf4_b64))) +vbool64_t vmseq(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16mf4_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m1_b32))) +vbool32_t vmseq(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m1_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m2_b16))) +vbool16_t vmseq(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m4_b8))) +vbool8_t vmseq(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m4_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m8_b4))) +vbool4_t vmseq(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m8_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32mf2_b64))) +vbool64_t vmseq(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32mf2_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m1_b64))) +vbool64_t vmseq(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m1_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m2_b32))) +vbool32_t vmseq(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m4_b16))) +vbool16_t vmseq(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m4_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m8_b8))) +vbool8_t vmseq(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m8_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m1_b8))) +vbool8_t vmseq(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m1_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m2_b4))) +vbool4_t vmseq(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m2_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m4_b2))) +vbool2_t vmseq(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m4_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m8_b1))) +vbool1_t vmseq(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m8_b1_m))) +vbool1_t vmseq(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8mf2_b16))) +vbool16_t vmseq(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8mf2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8mf4_b32))) +vbool32_t vmseq(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8mf4_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8mf8_b64))) +vbool64_t vmseq(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8mf8_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m1_b16))) +vbool16_t vmseq(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m1_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m2_b8))) +vbool8_t vmseq(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m2_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m4_b4))) +vbool4_t vmseq(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m4_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m8_b2))) +vbool2_t vmseq(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m8_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16mf2_b32))) +vbool32_t vmseq(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16mf2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16mf4_b64))) +vbool64_t vmseq(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16mf4_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m1_b32))) +vbool32_t vmseq(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m1_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m2_b16))) +vbool16_t vmseq(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m4_b8))) +vbool8_t vmseq(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m4_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m8_b4))) +vbool4_t vmseq(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m8_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32mf2_b64))) +vbool64_t vmseq(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32mf2_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m1_b64))) +vbool64_t vmseq(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m1_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m2_b32))) +vbool32_t vmseq(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m4_b16))) +vbool16_t vmseq(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m4_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m8_b8))) +vbool8_t vmseq(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m8_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m1_b8))) +vbool8_t vmsne(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m1_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m2_b4))) +vbool4_t vmsne(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m2_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m4_b2))) +vbool2_t vmsne(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m4_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m8_b1))) +vbool1_t vmsne(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m8_b1_m))) +vbool1_t vmsne(vbool1_t op0, vbool1_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8mf2_b16))) +vbool16_t vmsne(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8mf2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8mf4_b32))) +vbool32_t vmsne(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8mf4_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8mf8_b64))) +vbool64_t vmsne(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8mf8_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m1_b16))) +vbool16_t vmsne(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m1_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m2_b8))) +vbool8_t vmsne(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m2_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m4_b4))) +vbool4_t vmsne(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m4_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m8_b2))) +vbool2_t vmsne(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m8_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16mf2_b32))) +vbool32_t vmsne(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16mf2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16mf4_b64))) +vbool64_t vmsne(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16mf4_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m1_b32))) +vbool32_t vmsne(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m1_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m2_b16))) +vbool16_t vmsne(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m4_b8))) +vbool8_t vmsne(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m4_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m8_b4))) +vbool4_t vmsne(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m8_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32mf2_b64))) +vbool64_t vmsne(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32mf2_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m1_b64))) +vbool64_t vmsne(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m1_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m2_b32))) +vbool32_t vmsne(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m4_b16))) +vbool16_t vmsne(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m4_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m8_b8))) +vbool8_t vmsne(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m8_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m1_b8))) +vbool8_t vmsne(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m1_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m2_b4))) +vbool4_t vmsne(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m2_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m4_b2))) +vbool2_t vmsne(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m4_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m8_b1))) +vbool1_t vmsne(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m8_b1_m))) +vbool1_t vmsne(vbool1_t op0, vbool1_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8mf2_b16))) +vbool16_t vmsne(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8mf2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8mf4_b32))) +vbool32_t vmsne(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8mf4_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8mf8_b64))) +vbool64_t vmsne(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8mf8_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m1_b16))) +vbool16_t vmsne(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m1_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m2_b8))) +vbool8_t vmsne(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m2_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m4_b4))) +vbool4_t vmsne(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m4_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m8_b2))) +vbool2_t vmsne(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m8_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16mf2_b32))) +vbool32_t vmsne(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16mf2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16mf4_b64))) +vbool64_t vmsne(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16mf4_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m1_b32))) +vbool32_t vmsne(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m1_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m2_b16))) +vbool16_t vmsne(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m4_b8))) +vbool8_t vmsne(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m4_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m8_b4))) +vbool4_t vmsne(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m8_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32mf2_b64))) +vbool64_t vmsne(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32mf2_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m1_b64))) +vbool64_t vmsne(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m1_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m2_b32))) +vbool32_t vmsne(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m4_b16))) +vbool16_t vmsne(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m4_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m8_b8))) +vbool8_t vmsne(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m8_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m1_b8))) +vbool8_t vmsne(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m1_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m2_b4))) +vbool4_t vmsne(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m2_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m4_b2))) +vbool2_t vmsne(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m4_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m8_b1))) +vbool1_t vmsne(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m8_b1_m))) +vbool1_t vmsne(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8mf2_b16))) +vbool16_t vmsne(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8mf2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8mf4_b32))) +vbool32_t vmsne(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8mf4_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8mf8_b64))) +vbool64_t vmsne(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8mf8_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m1_b16))) +vbool16_t vmsne(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m1_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m2_b8))) +vbool8_t vmsne(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m2_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m4_b4))) +vbool4_t vmsne(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m4_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m8_b2))) +vbool2_t vmsne(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m8_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16mf2_b32))) +vbool32_t vmsne(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16mf2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16mf4_b64))) +vbool64_t vmsne(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16mf4_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m1_b32))) +vbool32_t vmsne(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m1_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m2_b16))) +vbool16_t vmsne(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m4_b8))) +vbool8_t vmsne(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m4_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m8_b4))) +vbool4_t vmsne(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m8_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32mf2_b64))) +vbool64_t vmsne(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32mf2_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m1_b64))) +vbool64_t vmsne(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m1_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m2_b32))) +vbool32_t vmsne(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m4_b16))) +vbool16_t vmsne(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m4_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m8_b8))) +vbool8_t vmsne(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m8_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m1_b8))) +vbool8_t vmsne(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m1_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m2_b4))) +vbool4_t vmsne(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m2_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m4_b2))) +vbool2_t vmsne(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m4_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m8_b1))) +vbool1_t vmsne(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m8_b1_m))) +vbool1_t vmsne(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8mf2_b16))) +vbool16_t vmsne(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8mf2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8mf4_b32))) +vbool32_t vmsne(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8mf4_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8mf8_b64))) +vbool64_t vmsne(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8mf8_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m1_b16))) +vbool16_t vmsne(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m1_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m2_b8))) +vbool8_t vmsne(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m2_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m4_b4))) +vbool4_t vmsne(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m4_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m8_b2))) +vbool2_t vmsne(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m8_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16mf2_b32))) +vbool32_t vmsne(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16mf2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16mf4_b64))) +vbool64_t vmsne(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16mf4_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m1_b32))) +vbool32_t vmsne(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m1_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m2_b16))) +vbool16_t vmsne(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m4_b8))) +vbool8_t vmsne(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m4_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m8_b4))) +vbool4_t vmsne(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m8_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32mf2_b64))) +vbool64_t vmsne(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32mf2_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m1_b64))) +vbool64_t vmsne(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m1_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m2_b32))) +vbool32_t vmsne(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m4_b16))) +vbool16_t vmsne(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m4_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m8_b8))) +vbool8_t vmsne(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m8_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m1_b8))) +vbool8_t vmsltu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m1_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m2_b4))) +vbool4_t vmsltu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m2_b4_m))) +vbool4_t vmsltu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m4_b2))) +vbool2_t vmsltu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m4_b2_m))) +vbool2_t vmsltu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m8_b1))) +vbool1_t vmsltu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m8_b1_m))) +vbool1_t vmsltu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8mf2_b16))) +vbool16_t vmsltu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8mf2_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8mf4_b32))) +vbool32_t vmsltu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8mf4_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8mf8_b64))) +vbool64_t vmsltu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8mf8_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m1_b16))) +vbool16_t vmsltu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m1_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m2_b8))) +vbool8_t vmsltu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m2_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m4_b4))) +vbool4_t vmsltu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m4_b4_m))) +vbool4_t vmsltu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m8_b2))) +vbool2_t vmsltu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m8_b2_m))) +vbool2_t vmsltu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16mf2_b32))) +vbool32_t vmsltu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16mf2_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16mf4_b64))) +vbool64_t vmsltu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16mf4_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m1_b32))) +vbool32_t vmsltu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m1_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m2_b16))) +vbool16_t vmsltu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m2_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m4_b8))) +vbool8_t vmsltu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m4_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m8_b4))) +vbool4_t vmsltu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m8_b4_m))) +vbool4_t vmsltu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32mf2_b64))) +vbool64_t vmsltu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32mf2_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m1_b64))) +vbool64_t vmsltu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m1_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m2_b32))) +vbool32_t vmsltu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m2_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m4_b16))) +vbool16_t vmsltu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m4_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m8_b8))) +vbool8_t vmsltu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m8_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m1_b8))) +vbool8_t vmsltu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m1_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m2_b4))) +vbool4_t vmsltu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m2_b4_m))) +vbool4_t vmsltu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m4_b2))) +vbool2_t vmsltu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m4_b2_m))) +vbool2_t vmsltu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m8_b1))) +vbool1_t vmsltu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m8_b1_m))) +vbool1_t vmsltu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8mf2_b16))) +vbool16_t vmsltu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8mf2_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8mf4_b32))) +vbool32_t vmsltu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8mf4_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8mf8_b64))) +vbool64_t vmsltu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8mf8_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m1_b16))) +vbool16_t vmsltu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m1_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m2_b8))) +vbool8_t vmsltu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m2_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m4_b4))) +vbool4_t vmsltu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m4_b4_m))) +vbool4_t vmsltu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m8_b2))) +vbool2_t vmsltu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m8_b2_m))) +vbool2_t vmsltu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16mf2_b32))) +vbool32_t vmsltu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16mf2_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16mf4_b64))) +vbool64_t vmsltu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16mf4_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m1_b32))) +vbool32_t vmsltu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m1_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m2_b16))) +vbool16_t vmsltu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m2_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m4_b8))) +vbool8_t vmsltu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m4_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m8_b4))) +vbool4_t vmsltu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m8_b4_m))) +vbool4_t vmsltu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32mf2_b64))) +vbool64_t vmsltu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32mf2_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m1_b64))) +vbool64_t vmsltu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m1_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m2_b32))) +vbool32_t vmsltu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m2_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m4_b16))) +vbool16_t vmsltu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m4_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m8_b8))) +vbool8_t vmsltu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m8_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m1_b8))) +vbool8_t vmslt(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m1_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m2_b4))) +vbool4_t vmslt(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m2_b4_m))) +vbool4_t vmslt(vbool4_t op0, vbool4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m4_b2))) +vbool2_t vmslt(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m4_b2_m))) +vbool2_t vmslt(vbool2_t op0, vbool2_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m8_b1))) +vbool1_t vmslt(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m8_b1_m))) +vbool1_t vmslt(vbool1_t op0, vbool1_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8mf2_b16))) +vbool16_t vmslt(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8mf2_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8mf4_b32))) +vbool32_t vmslt(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8mf4_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8mf8_b64))) +vbool64_t vmslt(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8mf8_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m1_b16))) +vbool16_t vmslt(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m1_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m2_b8))) +vbool8_t vmslt(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m2_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m4_b4))) +vbool4_t vmslt(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m4_b4_m))) +vbool4_t vmslt(vbool4_t op0, vbool4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m8_b2))) +vbool2_t vmslt(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m8_b2_m))) +vbool2_t vmslt(vbool2_t op0, vbool2_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16mf2_b32))) +vbool32_t vmslt(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16mf2_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16mf4_b64))) +vbool64_t vmslt(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16mf4_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m1_b32))) +vbool32_t vmslt(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m1_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m2_b16))) +vbool16_t vmslt(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m2_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m4_b8))) +vbool8_t vmslt(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m4_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m8_b4))) +vbool4_t vmslt(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m8_b4_m))) +vbool4_t vmslt(vbool4_t op0, vbool4_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32mf2_b64))) +vbool64_t vmslt(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32mf2_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m1_b64))) +vbool64_t vmslt(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m1_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m2_b32))) +vbool32_t vmslt(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m2_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m4_b16))) +vbool16_t vmslt(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m4_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m8_b8))) +vbool8_t vmslt(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m8_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16m1))) +void vsuxei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16m1_m))) +void vsuxei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16m2))) +void vsuxei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16m2_m))) +void vsuxei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16m4))) +void vsuxei32(uint16_t * op0, vuint32m8_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16m4_m))) +void vsuxei32(vbool4_t op0, uint16_t * op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16mf2))) +void vsuxei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16mf2_m))) +void vsuxei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16mf4))) +void vsuxei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16mf4_m))) +void vsuxei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m1_b8))) +vbool8_t vmslt(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m1_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m2_b4))) +vbool4_t vmslt(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m2_b4_m))) +vbool4_t vmslt(vbool4_t op0, vbool4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m4_b2))) +vbool2_t vmslt(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m4_b2_m))) +vbool2_t vmslt(vbool2_t op0, vbool2_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m8_b1))) +vbool1_t vmslt(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m8_b1_m))) +vbool1_t vmslt(vbool1_t op0, vbool1_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8mf2_b16))) +vbool16_t vmslt(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8mf2_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8mf4_b32))) +vbool32_t vmslt(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8mf4_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8mf8_b64))) +vbool64_t vmslt(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8mf8_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m1_b16))) +vbool16_t vmslt(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m1_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m2_b8))) +vbool8_t vmslt(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m2_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m4_b4))) +vbool4_t vmslt(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m4_b4_m))) +vbool4_t vmslt(vbool4_t op0, vbool4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m8_b2))) +vbool2_t vmslt(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m8_b2_m))) +vbool2_t vmslt(vbool2_t op0, vbool2_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16mf2_b32))) +vbool32_t vmslt(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16mf2_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16mf4_b64))) +vbool64_t vmslt(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16mf4_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m1_b32))) +vbool32_t vmslt(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m1_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m2_b16))) +vbool16_t vmslt(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m2_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m4_b8))) +vbool8_t vmslt(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m4_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m8_b4))) +vbool4_t vmslt(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m8_b4_m))) +vbool4_t vmslt(vbool4_t op0, vbool4_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32mf2_b64))) +vbool64_t vmslt(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32mf2_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m1_b64))) +vbool64_t vmslt(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m1_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m2_b32))) +vbool32_t vmslt(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m2_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m4_b16))) +vbool16_t vmslt(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m4_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m8_b8))) +vbool8_t vmslt(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m8_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m1_b8))) +vbool8_t vmsleu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m1_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m2_b4))) +vbool4_t vmsleu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m2_b4_m))) +vbool4_t vmsleu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m4_b2))) +vbool2_t vmsleu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m4_b2_m))) +vbool2_t vmsleu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m8_b1))) +vbool1_t vmsleu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m8_b1_m))) +vbool1_t vmsleu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8mf2_b16))) +vbool16_t vmsleu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8mf2_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8mf4_b32))) +vbool32_t vmsleu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8mf4_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8mf8_b64))) +vbool64_t vmsleu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8mf8_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m1_b16))) +vbool16_t vmsleu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m1_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m2_b8))) +vbool8_t vmsleu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m2_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m4_b4))) +vbool4_t vmsleu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m4_b4_m))) +vbool4_t vmsleu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m8_b2))) +vbool2_t vmsleu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m8_b2_m))) +vbool2_t vmsleu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16mf2_b32))) +vbool32_t vmsleu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16mf2_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16mf4_b64))) +vbool64_t vmsleu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16mf4_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m1_b32))) +vbool32_t vmsleu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m1_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m2_b16))) +vbool16_t vmsleu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m2_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m4_b8))) +vbool8_t vmsleu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m4_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m8_b4))) +vbool4_t vmsleu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m8_b4_m))) +vbool4_t vmsleu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32mf2_b64))) +vbool64_t vmsleu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32mf2_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m1_b64))) +vbool64_t vmsleu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m1_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m2_b32))) +vbool32_t vmsleu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m2_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m4_b16))) +vbool16_t vmsleu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m4_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m8_b8))) +vbool8_t vmsleu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m8_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m1_b8))) +vbool8_t vmsleu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m1_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m2_b4))) +vbool4_t vmsleu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m2_b4_m))) +vbool4_t vmsleu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m4_b2))) +vbool2_t vmsleu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m4_b2_m))) +vbool2_t vmsleu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m8_b1))) +vbool1_t vmsleu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m8_b1_m))) +vbool1_t vmsleu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8mf2_b16))) +vbool16_t vmsleu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8mf2_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8mf4_b32))) +vbool32_t vmsleu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8mf4_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8mf8_b64))) +vbool64_t vmsleu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8mf8_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m1_b16))) +vbool16_t vmsleu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m1_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m2_b8))) +vbool8_t vmsleu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m2_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m4_b4))) +vbool4_t vmsleu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m4_b4_m))) +vbool4_t vmsleu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m8_b2))) +vbool2_t vmsleu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m8_b2_m))) +vbool2_t vmsleu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16mf2_b32))) +vbool32_t vmsleu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16mf2_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16mf4_b64))) +vbool64_t vmsleu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16mf4_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m1_b32))) +vbool32_t vmsleu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m1_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m2_b16))) +vbool16_t vmsleu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m2_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m4_b8))) +vbool8_t vmsleu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m4_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m8_b4))) +vbool4_t vmsleu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m8_b4_m))) +vbool4_t vmsleu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32mf2_b64))) +vbool64_t vmsleu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32mf2_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m1_b64))) +vbool64_t vmsleu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m1_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m2_b32))) +vbool32_t vmsleu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m2_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m4_b16))) +vbool16_t vmsleu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m4_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m8_b8))) +vbool8_t vmsleu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m8_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m1_b8))) +vbool8_t vmsle(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m1_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m2_b4))) +vbool4_t vmsle(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m2_b4_m))) +vbool4_t vmsle(vbool4_t op0, vbool4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m4_b2))) +vbool2_t vmsle(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m4_b2_m))) +vbool2_t vmsle(vbool2_t op0, vbool2_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m8_b1))) +vbool1_t vmsle(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m8_b1_m))) +vbool1_t vmsle(vbool1_t op0, vbool1_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8mf2_b16))) +vbool16_t vmsle(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8mf2_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8mf4_b32))) +vbool32_t vmsle(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8mf4_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8mf8_b64))) +vbool64_t vmsle(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8mf8_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m1_b16))) +vbool16_t vmsle(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m1_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m2_b8))) +vbool8_t vmsle(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m2_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m4_b4))) +vbool4_t vmsle(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m4_b4_m))) +vbool4_t vmsle(vbool4_t op0, vbool4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m8_b2))) +vbool2_t vmsle(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m8_b2_m))) +vbool2_t vmsle(vbool2_t op0, vbool2_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16mf2_b32))) +vbool32_t vmsle(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16mf2_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16mf4_b64))) +vbool64_t vmsle(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16mf4_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m1_b32))) +vbool32_t vmsle(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m1_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m2_b16))) +vbool16_t vmsle(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m2_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m4_b8))) +vbool8_t vmsle(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m4_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m8_b4))) +vbool4_t vmsle(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m8_b4_m))) +vbool4_t vmsle(vbool4_t op0, vbool4_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32mf2_b64))) +vbool64_t vmsle(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32mf2_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m1_b64))) +vbool64_t vmsle(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m1_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m2_b32))) +vbool32_t vmsle(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m2_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m4_b16))) +vbool16_t vmsle(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m4_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m8_b8))) +vbool8_t vmsle(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m8_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m1_b8))) +vbool8_t vmsle(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m1_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m2_b4))) +vbool4_t vmsle(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m2_b4_m))) +vbool4_t vmsle(vbool4_t op0, vbool4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m4_b2))) +vbool2_t vmsle(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m4_b2_m))) +vbool2_t vmsle(vbool2_t op0, vbool2_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m8_b1))) +vbool1_t vmsle(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m8_b1_m))) +vbool1_t vmsle(vbool1_t op0, vbool1_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8mf2_b16))) +vbool16_t vmsle(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8mf2_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8mf4_b32))) +vbool32_t vmsle(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8mf4_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8mf8_b64))) +vbool64_t vmsle(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8mf8_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m1_b16))) +vbool16_t vmsle(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m1_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m2_b8))) +vbool8_t vmsle(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m2_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m4_b4))) +vbool4_t vmsle(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m4_b4_m))) +vbool4_t vmsle(vbool4_t op0, vbool4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m8_b2))) +vbool2_t vmsle(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m8_b2_m))) +vbool2_t vmsle(vbool2_t op0, vbool2_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16mf2_b32))) +vbool32_t vmsle(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16mf2_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16mf4_b64))) +vbool64_t vmsle(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16mf4_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m1_b32))) +vbool32_t vmsle(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m1_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m2_b16))) +vbool16_t vmsle(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m2_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m4_b8))) +vbool8_t vmsle(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m4_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m8_b4))) +vbool4_t vmsle(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m8_b4_m))) +vbool4_t vmsle(vbool4_t op0, vbool4_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32mf2_b64))) +vbool64_t vmsle(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32mf2_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m1_b64))) +vbool64_t vmsle(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m1_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m2_b32))) +vbool32_t vmsle(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m2_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m4_b16))) +vbool16_t vmsle(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m4_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m8_b8))) +vbool8_t vmsle(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m8_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m1_b8))) +vbool8_t vmsgtu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m1_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m2_b4))) +vbool4_t vmsgtu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m2_b4_m))) +vbool4_t vmsgtu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m4_b2))) +vbool2_t vmsgtu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m4_b2_m))) +vbool2_t vmsgtu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m8_b1))) +vbool1_t vmsgtu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m8_b1_m))) +vbool1_t vmsgtu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8mf2_b16))) +vbool16_t vmsgtu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8mf2_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8mf4_b32))) +vbool32_t vmsgtu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8mf4_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8mf8_b64))) +vbool64_t vmsgtu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8mf8_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m1_b16))) +vbool16_t vmsgtu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m1_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m2_b8))) +vbool8_t vmsgtu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m2_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m4_b4))) +vbool4_t vmsgtu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m4_b4_m))) +vbool4_t vmsgtu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m8_b2))) +vbool2_t vmsgtu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m8_b2_m))) +vbool2_t vmsgtu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16mf2_b32))) +vbool32_t vmsgtu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16mf2_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16mf4_b64))) +vbool64_t vmsgtu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16mf4_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m1_b32))) +vbool32_t vmsgtu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m1_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m2_b16))) +vbool16_t vmsgtu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m2_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m4_b8))) +vbool8_t vmsgtu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m4_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m8_b4))) +vbool4_t vmsgtu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m8_b4_m))) +vbool4_t vmsgtu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32mf2_b64))) +vbool64_t vmsgtu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32mf2_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m1_b64))) +vbool64_t vmsgtu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m1_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m2_b32))) +vbool32_t vmsgtu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m2_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m4_b16))) +vbool16_t vmsgtu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m4_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m8_b8))) +vbool8_t vmsgtu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m8_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m1_b8))) +vbool8_t vmsgtu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m1_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m2_b4))) +vbool4_t vmsgtu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m2_b4_m))) +vbool4_t vmsgtu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m4_b2))) +vbool2_t vmsgtu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m4_b2_m))) +vbool2_t vmsgtu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m8_b1))) +vbool1_t vmsgtu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m8_b1_m))) +vbool1_t vmsgtu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8mf2_b16))) +vbool16_t vmsgtu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8mf2_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8mf4_b32))) +vbool32_t vmsgtu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8mf4_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8mf8_b64))) +vbool64_t vmsgtu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8mf8_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m1_b16))) +vbool16_t vmsgtu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m1_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m2_b8))) +vbool8_t vmsgtu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m2_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m4_b4))) +vbool4_t vmsgtu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m4_b4_m))) +vbool4_t vmsgtu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m8_b2))) +vbool2_t vmsgtu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m8_b2_m))) +vbool2_t vmsgtu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16mf2_b32))) +vbool32_t vmsgtu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16mf2_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16mf4_b64))) +vbool64_t vmsgtu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16mf4_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m1_b32))) +vbool32_t vmsgtu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m1_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m2_b16))) +vbool16_t vmsgtu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m2_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m4_b8))) +vbool8_t vmsgtu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m4_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m8_b4))) +vbool4_t vmsgtu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m8_b4_m))) +vbool4_t vmsgtu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32mf2_b64))) +vbool64_t vmsgtu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32mf2_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m1_b64))) +vbool64_t vmsgtu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m1_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m2_b32))) +vbool32_t vmsgtu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m2_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m4_b16))) +vbool16_t vmsgtu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m4_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m8_b8))) +vbool8_t vmsgtu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m8_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m1_b8))) +vbool8_t vmsgt(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m1_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m2_b4))) +vbool4_t vmsgt(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m2_b4_m))) +vbool4_t vmsgt(vbool4_t op0, vbool4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m4_b2))) +vbool2_t vmsgt(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m4_b2_m))) +vbool2_t vmsgt(vbool2_t op0, vbool2_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m8_b1))) +vbool1_t vmsgt(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m8_b1_m))) +vbool1_t vmsgt(vbool1_t op0, vbool1_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8mf2_b16))) +vbool16_t vmsgt(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8mf2_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8mf4_b32))) +vbool32_t vmsgt(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8mf4_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8mf8_b64))) +vbool64_t vmsgt(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8mf8_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m1_b16))) +vbool16_t vmsgt(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m1_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m2_b8))) +vbool8_t vmsgt(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m2_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m4_b4))) +vbool4_t vmsgt(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m4_b4_m))) +vbool4_t vmsgt(vbool4_t op0, vbool4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m8_b2))) +vbool2_t vmsgt(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m8_b2_m))) +vbool2_t vmsgt(vbool2_t op0, vbool2_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16mf2_b32))) +vbool32_t vmsgt(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16mf2_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16mf4_b64))) +vbool64_t vmsgt(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16mf4_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m1_b32))) +vbool32_t vmsgt(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m1_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m2_b16))) +vbool16_t vmsgt(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m2_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m4_b8))) +vbool8_t vmsgt(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m4_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m8_b4))) +vbool4_t vmsgt(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m8_b4_m))) +vbool4_t vmsgt(vbool4_t op0, vbool4_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32mf2_b64))) +vbool64_t vmsgt(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32mf2_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m1_b64))) +vbool64_t vmsgt(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m1_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m2_b32))) +vbool32_t vmsgt(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m2_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m4_b16))) +vbool16_t vmsgt(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m4_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m8_b8))) +vbool8_t vmsgt(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m8_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m1_b8))) +vbool8_t vmsgt(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m1_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m2_b4))) +vbool4_t vmsgt(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m2_b4_m))) +vbool4_t vmsgt(vbool4_t op0, vbool4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m4_b2))) +vbool2_t vmsgt(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m4_b2_m))) +vbool2_t vmsgt(vbool2_t op0, vbool2_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m8_b1))) +vbool1_t vmsgt(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m8_b1_m))) +vbool1_t vmsgt(vbool1_t op0, vbool1_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8mf2_b16))) +vbool16_t vmsgt(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8mf2_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8mf4_b32))) +vbool32_t vmsgt(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8mf4_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8mf8_b64))) +vbool64_t vmsgt(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8mf8_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m1_b16))) +vbool16_t vmsgt(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m1_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m2_b8))) +vbool8_t vmsgt(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m2_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m4_b4))) +vbool4_t vmsgt(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m4_b4_m))) +vbool4_t vmsgt(vbool4_t op0, vbool4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m8_b2))) +vbool2_t vmsgt(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m8_b2_m))) +vbool2_t vmsgt(vbool2_t op0, vbool2_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16mf2_b32))) +vbool32_t vmsgt(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16mf2_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16mf4_b64))) +vbool64_t vmsgt(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16mf4_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m1_b32))) +vbool32_t vmsgt(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m1_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m2_b16))) +vbool16_t vmsgt(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m2_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m4_b8))) +vbool8_t vmsgt(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m4_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m8_b4))) +vbool4_t vmsgt(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m8_b4_m))) +vbool4_t vmsgt(vbool4_t op0, vbool4_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32mf2_b64))) +vbool64_t vmsgt(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32mf2_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m1_b64))) +vbool64_t vmsgt(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m1_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m2_b32))) +vbool32_t vmsgt(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m2_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m4_b16))) +vbool16_t vmsgt(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m4_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m8_b8))) +vbool8_t vmsgt(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m8_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m1_b8))) +vbool8_t vmsgeu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m1_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m2_b4))) +vbool4_t vmsgeu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m2_b4_m))) +vbool4_t vmsgeu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m4_b2))) +vbool2_t vmsgeu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m4_b2_m))) +vbool2_t vmsgeu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m8_b1))) +vbool1_t vmsgeu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m8_b1_m))) +vbool1_t vmsgeu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8mf2_b16))) +vbool16_t vmsgeu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8mf2_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8mf4_b32))) +vbool32_t vmsgeu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8mf4_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8mf8_b64))) +vbool64_t vmsgeu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8mf8_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m1_b16))) +vbool16_t vmsgeu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m1_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m2_b8))) +vbool8_t vmsgeu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m2_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m4_b4))) +vbool4_t vmsgeu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m4_b4_m))) +vbool4_t vmsgeu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m8_b2))) +vbool2_t vmsgeu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m8_b2_m))) +vbool2_t vmsgeu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16mf2_b32))) +vbool32_t vmsgeu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16mf2_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16mf4_b64))) +vbool64_t vmsgeu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16mf4_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m1_b32))) +vbool32_t vmsgeu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m1_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m2_b16))) +vbool16_t vmsgeu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m2_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m4_b8))) +vbool8_t vmsgeu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m4_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m8_b4))) +vbool4_t vmsgeu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m8_b4_m))) +vbool4_t vmsgeu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32mf2_b64))) +vbool64_t vmsgeu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32mf2_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m1_b64))) +vbool64_t vmsgeu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m1_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m2_b32))) +vbool32_t vmsgeu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m2_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m4_b16))) +vbool16_t vmsgeu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m4_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m8_b8))) +vbool8_t vmsgeu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m8_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16m1))) +void vsuxei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16m1_m))) +void vsuxei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16m2))) +void vsuxei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16m2_m))) +void vsuxei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16mf2))) +void vsuxei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16mf2_m))) +void vsuxei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16mf4))) +void vsuxei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16mf4_m))) +void vsuxei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m1_b8))) +vbool8_t vmsgeu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m1_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m2_b4))) +vbool4_t vmsgeu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m2_b4_m))) +vbool4_t vmsgeu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m4_b2))) +vbool2_t vmsgeu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m4_b2_m))) +vbool2_t vmsgeu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m8_b1))) +vbool1_t vmsgeu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m8_b1_m))) +vbool1_t vmsgeu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8mf2_b16))) +vbool16_t vmsgeu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8mf2_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8mf4_b32))) +vbool32_t vmsgeu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8mf4_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8mf8_b64))) +vbool64_t vmsgeu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8mf8_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m1_b16))) +vbool16_t vmsgeu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m1_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m2_b8))) +vbool8_t vmsgeu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m2_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m4_b4))) +vbool4_t vmsgeu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m4_b4_m))) +vbool4_t vmsgeu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m8_b2))) +vbool2_t vmsgeu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m8_b2_m))) +vbool2_t vmsgeu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16mf2_b32))) +vbool32_t vmsgeu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16mf2_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16mf4_b64))) +vbool64_t vmsgeu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16mf4_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m1_b32))) +vbool32_t vmsgeu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m1_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m2_b16))) +vbool16_t vmsgeu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m2_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m4_b8))) +vbool8_t vmsgeu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m4_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m8_b4))) +vbool4_t vmsgeu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m8_b4_m))) +vbool4_t vmsgeu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32mf2_b64))) +vbool64_t vmsgeu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32mf2_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m1_b64))) +vbool64_t vmsgeu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m1_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m2_b32))) +vbool32_t vmsgeu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m2_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m4_b16))) +vbool16_t vmsgeu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m4_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m8_b8))) +vbool8_t vmsgeu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m8_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m1_b8))) +vbool8_t vmsge(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m1_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m2_b4))) +vbool4_t vmsge(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m2_b4_m))) +vbool4_t vmsge(vbool4_t op0, vbool4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m4_b2))) +vbool2_t vmsge(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m4_b2_m))) +vbool2_t vmsge(vbool2_t op0, vbool2_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m8_b1))) +vbool1_t vmsge(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m8_b1_m))) +vbool1_t vmsge(vbool1_t op0, vbool1_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8mf2_b16))) +vbool16_t vmsge(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8mf2_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8mf4_b32))) +vbool32_t vmsge(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8mf4_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8mf8_b64))) +vbool64_t vmsge(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8mf8_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m1_b16))) +vbool16_t vmsge(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m1_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m2_b8))) +vbool8_t vmsge(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m2_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m4_b4))) +vbool4_t vmsge(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m4_b4_m))) +vbool4_t vmsge(vbool4_t op0, vbool4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m8_b2))) +vbool2_t vmsge(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m8_b2_m))) +vbool2_t vmsge(vbool2_t op0, vbool2_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16mf2_b32))) +vbool32_t vmsge(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16mf2_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16mf4_b64))) +vbool64_t vmsge(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16mf4_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m1_b32))) +vbool32_t vmsge(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m1_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m2_b16))) +vbool16_t vmsge(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m2_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m4_b8))) +vbool8_t vmsge(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m4_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m8_b4))) +vbool4_t vmsge(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m8_b4_m))) +vbool4_t vmsge(vbool4_t op0, vbool4_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32mf2_b64))) +vbool64_t vmsge(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32mf2_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m1_b64))) +vbool64_t vmsge(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m1_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m2_b32))) +vbool32_t vmsge(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m2_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m4_b16))) +vbool16_t vmsge(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m4_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m8_b8))) +vbool8_t vmsge(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m8_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m1_b8))) +vbool8_t vmsge(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m1_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m2_b4))) +vbool4_t vmsge(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m2_b4_m))) +vbool4_t vmsge(vbool4_t op0, vbool4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m4_b2))) +vbool2_t vmsge(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m4_b2_m))) +vbool2_t vmsge(vbool2_t op0, vbool2_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m8_b1))) +vbool1_t vmsge(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m8_b1_m))) +vbool1_t vmsge(vbool1_t op0, vbool1_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8mf2_b16))) +vbool16_t vmsge(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8mf2_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8mf4_b32))) +vbool32_t vmsge(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8mf4_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8mf8_b64))) +vbool64_t vmsge(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8mf8_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m1_b16))) +vbool16_t vmsge(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m1_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m2_b8))) +vbool8_t vmsge(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m2_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m4_b4))) +vbool4_t vmsge(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m4_b4_m))) +vbool4_t vmsge(vbool4_t op0, vbool4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m8_b2))) +vbool2_t vmsge(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m8_b2_m))) +vbool2_t vmsge(vbool2_t op0, vbool2_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16mf2_b32))) +vbool32_t vmsge(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16mf2_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16mf4_b64))) +vbool64_t vmsge(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16mf4_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m1_b32))) +vbool32_t vmsge(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m1_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m2_b16))) +vbool16_t vmsge(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m2_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m4_b8))) +vbool8_t vmsge(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m4_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m8_b4))) +vbool4_t vmsge(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m8_b4_m))) +vbool4_t vmsge(vbool4_t op0, vbool4_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32mf2_b64))) +vbool64_t vmsge(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32mf2_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m1_b64))) +vbool64_t vmsge(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m1_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m2_b32))) +vbool32_t vmsge(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m2_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m4_b16))) +vbool16_t vmsge(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m4_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m8_b8))) +vbool8_t vmsge(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m8_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m1))) +vuint8m1_t vminu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m1_m))) +vuint8m1_t vminu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m2))) +vuint8m2_t vminu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m2_m))) +vuint8m2_t vminu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m4))) +vuint8m4_t vminu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m4_m))) +vuint8m4_t vminu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m8))) +vuint8m8_t vminu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m8_m))) +vuint8m8_t vminu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8mf2))) +vuint8mf2_t vminu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8mf2_m))) +vuint8mf2_t vminu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8mf4))) +vuint8mf4_t vminu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8mf4_m))) +vuint8mf4_t vminu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8mf8))) +vuint8mf8_t vminu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8mf8_m))) +vuint8mf8_t vminu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m1))) +vuint16m1_t vminu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m1_m))) +vuint16m1_t vminu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m2))) +vuint16m2_t vminu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m2_m))) +vuint16m2_t vminu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m4))) +vuint16m4_t vminu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m4_m))) +vuint16m4_t vminu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m8))) +vuint16m8_t vminu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m8_m))) +vuint16m8_t vminu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16mf2))) +vuint16mf2_t vminu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16mf2_m))) +vuint16mf2_t vminu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16mf4))) +vuint16mf4_t vminu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16mf4_m))) +vuint16mf4_t vminu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m1))) +vuint32m1_t vminu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m1_m))) +vuint32m1_t vminu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m2))) +vuint32m2_t vminu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m2_m))) +vuint32m2_t vminu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m4))) +vuint32m4_t vminu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m4_m))) +vuint32m4_t vminu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m8))) +vuint32m8_t vminu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m8_m))) +vuint32m8_t vminu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32mf2))) +vuint32mf2_t vminu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32mf2_m))) +vuint32mf2_t vminu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m1))) +vuint64m1_t vminu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m1_m))) +vuint64m1_t vminu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m2))) +vuint64m2_t vminu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m2_m))) +vuint64m2_t vminu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m4))) +vuint64m4_t vminu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m4_m))) +vuint64m4_t vminu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m8))) +vuint64m8_t vminu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m8_m))) +vuint64m8_t vminu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m1))) +vuint8m1_t vminu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m1_m))) +vuint8m1_t vminu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m2))) +vuint8m2_t vminu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m2_m))) +vuint8m2_t vminu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m4))) +vuint8m4_t vminu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m4_m))) +vuint8m4_t vminu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m8))) +vuint8m8_t vminu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m8_m))) +vuint8m8_t vminu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8mf2))) +vuint8mf2_t vminu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8mf2_m))) +vuint8mf2_t vminu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8mf4))) +vuint8mf4_t vminu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8mf4_m))) +vuint8mf4_t vminu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8mf8))) +vuint8mf8_t vminu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8mf8_m))) +vuint8mf8_t vminu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m1))) +vuint16m1_t vminu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m1_m))) +vuint16m1_t vminu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m2))) +vuint16m2_t vminu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m2_m))) +vuint16m2_t vminu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m4))) +vuint16m4_t vminu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m4_m))) +vuint16m4_t vminu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m8))) +vuint16m8_t vminu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m8_m))) +vuint16m8_t vminu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16mf2))) +vuint16mf2_t vminu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16mf2_m))) +vuint16mf2_t vminu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16mf4))) +vuint16mf4_t vminu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16mf4_m))) +vuint16mf4_t vminu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m1))) +vuint32m1_t vminu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m1_m))) +vuint32m1_t vminu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m2))) +vuint32m2_t vminu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m2_m))) +vuint32m2_t vminu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m4))) +vuint32m4_t vminu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m4_m))) +vuint32m4_t vminu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m8))) +vuint32m8_t vminu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m8_m))) +vuint32m8_t vminu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32mf2))) +vuint32mf2_t vminu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32mf2_m))) +vuint32mf2_t vminu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m1))) +vuint64m1_t vminu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m1_m))) +vuint64m1_t vminu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m2))) +vuint64m2_t vminu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m2_m))) +vuint64m2_t vminu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m4))) +vuint64m4_t vminu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m4_m))) +vuint64m4_t vminu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m8))) +vuint64m8_t vminu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m8_m))) +vuint64m8_t vminu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m1))) +vint8m1_t vmin(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m1_m))) +vint8m1_t vmin(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m2))) +vint8m2_t vmin(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m2_m))) +vint8m2_t vmin(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m4))) +vint8m4_t vmin(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m4_m))) +vint8m4_t vmin(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m8))) +vint8m8_t vmin(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m8_m))) +vint8m8_t vmin(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8mf2))) +vint8mf2_t vmin(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8mf2_m))) +vint8mf2_t vmin(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8mf4))) +vint8mf4_t vmin(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8mf4_m))) +vint8mf4_t vmin(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8mf8))) +vint8mf8_t vmin(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8mf8_m))) +vint8mf8_t vmin(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m1))) +vint16m1_t vmin(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m1_m))) +vint16m1_t vmin(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m2))) +vint16m2_t vmin(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m2_m))) +vint16m2_t vmin(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m4))) +vint16m4_t vmin(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m4_m))) +vint16m4_t vmin(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m8))) +vint16m8_t vmin(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m8_m))) +vint16m8_t vmin(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16mf2))) +vint16mf2_t vmin(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16mf2_m))) +vint16mf2_t vmin(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16mf4))) +vint16mf4_t vmin(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16mf4_m))) +vint16mf4_t vmin(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m1))) +vint32m1_t vmin(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m1_m))) +vint32m1_t vmin(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m2))) +vint32m2_t vmin(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m2_m))) +vint32m2_t vmin(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m4))) +vint32m4_t vmin(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m4_m))) +vint32m4_t vmin(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m8))) +vint32m8_t vmin(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m8_m))) +vint32m8_t vmin(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32mf2))) +vint32mf2_t vmin(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32mf2_m))) +vint32mf2_t vmin(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m1))) +vint64m1_t vmin(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m1_m))) +vint64m1_t vmin(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m2))) +vint64m2_t vmin(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m2_m))) +vint64m2_t vmin(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m4))) +vint64m4_t vmin(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m4_m))) +vint64m4_t vmin(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m8))) +vint64m8_t vmin(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m8_m))) +vint64m8_t vmin(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m1))) +vint8m1_t vmin(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m1_m))) +vint8m1_t vmin(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m2))) +vint8m2_t vmin(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m2_m))) +vint8m2_t vmin(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m4))) +vint8m4_t vmin(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m4_m))) +vint8m4_t vmin(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m8))) +vint8m8_t vmin(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m8_m))) +vint8m8_t vmin(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8mf2))) +vint8mf2_t vmin(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8mf2_m))) +vint8mf2_t vmin(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8mf4))) +vint8mf4_t vmin(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8mf4_m))) +vint8mf4_t vmin(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8mf8))) +vint8mf8_t vmin(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8mf8_m))) +vint8mf8_t vmin(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m1))) +vint16m1_t vmin(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m1_m))) +vint16m1_t vmin(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m2))) +vint16m2_t vmin(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m2_m))) +vint16m2_t vmin(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m4))) +vint16m4_t vmin(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m4_m))) +vint16m4_t vmin(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m8))) +vint16m8_t vmin(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m8_m))) +vint16m8_t vmin(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16mf2))) +vint16mf2_t vmin(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16mf2_m))) +vint16mf2_t vmin(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16mf4))) +vint16mf4_t vmin(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16mf4_m))) +vint16mf4_t vmin(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m1))) +vint32m1_t vmin(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m1_m))) +vint32m1_t vmin(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m2))) +vint32m2_t vmin(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m2_m))) +vint32m2_t vmin(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m4))) +vint32m4_t vmin(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m4_m))) +vint32m4_t vmin(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m8))) +vint32m8_t vmin(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m8_m))) +vint32m8_t vmin(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32mf2))) +vint32mf2_t vmin(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32mf2_m))) +vint32mf2_t vmin(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m1))) +vint64m1_t vmin(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m1_m))) +vint64m1_t vmin(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m2))) +vint64m2_t vmin(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m2_m))) +vint64m2_t vmin(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m4))) +vint64m4_t vmin(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m4_m))) +vint64m4_t vmin(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m8))) +vint64m8_t vmin(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m8_m))) +vint64m8_t vmin(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m1))) +vuint8m1_t vmaxu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m1_m))) +vuint8m1_t vmaxu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m2))) +vuint8m2_t vmaxu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m2_m))) +vuint8m2_t vmaxu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m4))) +vuint8m4_t vmaxu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m4_m))) +vuint8m4_t vmaxu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m8))) +vuint8m8_t vmaxu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m8_m))) +vuint8m8_t vmaxu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8mf2))) +vuint8mf2_t vmaxu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8mf2_m))) +vuint8mf2_t vmaxu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8mf4))) +vuint8mf4_t vmaxu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8mf4_m))) +vuint8mf4_t vmaxu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8mf8))) +vuint8mf8_t vmaxu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8mf8_m))) +vuint8mf8_t vmaxu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m1))) +vuint16m1_t vmaxu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m1_m))) +vuint16m1_t vmaxu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m2))) +vuint16m2_t vmaxu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m2_m))) +vuint16m2_t vmaxu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m4))) +vuint16m4_t vmaxu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m4_m))) +vuint16m4_t vmaxu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m8))) +vuint16m8_t vmaxu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m8_m))) +vuint16m8_t vmaxu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16mf2))) +vuint16mf2_t vmaxu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16mf2_m))) +vuint16mf2_t vmaxu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16mf4))) +vuint16mf4_t vmaxu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16mf4_m))) +vuint16mf4_t vmaxu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m1))) +vuint32m1_t vmaxu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m1_m))) +vuint32m1_t vmaxu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m2))) +vuint32m2_t vmaxu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m2_m))) +vuint32m2_t vmaxu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m4))) +vuint32m4_t vmaxu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m4_m))) +vuint32m4_t vmaxu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m8))) +vuint32m8_t vmaxu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m8_m))) +vuint32m8_t vmaxu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32mf2))) +vuint32mf2_t vmaxu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32mf2_m))) +vuint32mf2_t vmaxu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m1))) +vuint64m1_t vmaxu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m1_m))) +vuint64m1_t vmaxu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m2))) +vuint64m2_t vmaxu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m2_m))) +vuint64m2_t vmaxu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m4))) +vuint64m4_t vmaxu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m4_m))) +vuint64m4_t vmaxu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m8))) +vuint64m8_t vmaxu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m8_m))) +vuint64m8_t vmaxu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m1))) +vuint8m1_t vmaxu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m1_m))) +vuint8m1_t vmaxu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m2))) +vuint8m2_t vmaxu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m2_m))) +vuint8m2_t vmaxu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m4))) +vuint8m4_t vmaxu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m4_m))) +vuint8m4_t vmaxu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m8))) +vuint8m8_t vmaxu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m8_m))) +vuint8m8_t vmaxu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8mf2))) +vuint8mf2_t vmaxu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8mf2_m))) +vuint8mf2_t vmaxu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8mf4))) +vuint8mf4_t vmaxu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8mf4_m))) +vuint8mf4_t vmaxu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8mf8))) +vuint8mf8_t vmaxu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8mf8_m))) +vuint8mf8_t vmaxu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m1))) +vuint16m1_t vmaxu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m1_m))) +vuint16m1_t vmaxu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m2))) +vuint16m2_t vmaxu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m2_m))) +vuint16m2_t vmaxu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m4))) +vuint16m4_t vmaxu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m4_m))) +vuint16m4_t vmaxu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m8))) +vuint16m8_t vmaxu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m8_m))) +vuint16m8_t vmaxu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16mf2))) +vuint16mf2_t vmaxu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16mf2_m))) +vuint16mf2_t vmaxu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16mf4))) +vuint16mf4_t vmaxu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16mf4_m))) +vuint16mf4_t vmaxu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m1))) +vuint32m1_t vmaxu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m1_m))) +vuint32m1_t vmaxu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m2))) +vuint32m2_t vmaxu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m2_m))) +vuint32m2_t vmaxu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m4))) +vuint32m4_t vmaxu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m4_m))) +vuint32m4_t vmaxu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m8))) +vuint32m8_t vmaxu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m8_m))) +vuint32m8_t vmaxu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32mf2))) +vuint32mf2_t vmaxu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32mf2_m))) +vuint32mf2_t vmaxu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m1))) +vuint64m1_t vmaxu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m1_m))) +vuint64m1_t vmaxu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m2))) +vuint64m2_t vmaxu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m2_m))) +vuint64m2_t vmaxu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m4))) +vuint64m4_t vmaxu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m4_m))) +vuint64m4_t vmaxu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m8))) +vuint64m8_t vmaxu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m8_m))) +vuint64m8_t vmaxu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m1))) +vint8m1_t vmax(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m1_m))) +vint8m1_t vmax(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m2))) +vint8m2_t vmax(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m2_m))) +vint8m2_t vmax(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m4))) +vint8m4_t vmax(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m4_m))) +vint8m4_t vmax(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m8))) +vint8m8_t vmax(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m8_m))) +vint8m8_t vmax(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8mf2))) +vint8mf2_t vmax(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8mf2_m))) +vint8mf2_t vmax(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8mf4))) +vint8mf4_t vmax(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8mf4_m))) +vint8mf4_t vmax(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8mf8))) +vint8mf8_t vmax(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8mf8_m))) +vint8mf8_t vmax(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m1))) +vint16m1_t vmax(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m1_m))) +vint16m1_t vmax(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m2))) +vint16m2_t vmax(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m2_m))) +vint16m2_t vmax(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m4))) +vint16m4_t vmax(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m4_m))) +vint16m4_t vmax(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m8))) +vint16m8_t vmax(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m8_m))) +vint16m8_t vmax(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16mf2))) +vint16mf2_t vmax(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16mf2_m))) +vint16mf2_t vmax(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16mf4))) +vint16mf4_t vmax(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16mf4_m))) +vint16mf4_t vmax(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m1))) +vint32m1_t vmax(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m1_m))) +vint32m1_t vmax(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m2))) +vint32m2_t vmax(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m2_m))) +vint32m2_t vmax(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m4))) +vint32m4_t vmax(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m4_m))) +vint32m4_t vmax(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m8))) +vint32m8_t vmax(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m8_m))) +vint32m8_t vmax(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32mf2))) +vint32mf2_t vmax(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32mf2_m))) +vint32mf2_t vmax(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m1))) +vint64m1_t vmax(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m1_m))) +vint64m1_t vmax(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m2))) +vint64m2_t vmax(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m2_m))) +vint64m2_t vmax(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m4))) +vint64m4_t vmax(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m4_m))) +vint64m4_t vmax(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m8))) +vint64m8_t vmax(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m8_m))) +vint64m8_t vmax(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16m1))) +void vsuxei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16m1_m))) +void vsuxei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16m2))) +void vsuxei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16m2_m))) +void vsuxei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16mf2))) +void vsuxei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16mf2_m))) +void vsuxei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16mf4))) +void vsuxei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16mf4_m))) +void vsuxei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m1))) +vint8m1_t vmax(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m1_m))) +vint8m1_t vmax(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m2))) +vint8m2_t vmax(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m2_m))) +vint8m2_t vmax(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m4))) +vint8m4_t vmax(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m4_m))) +vint8m4_t vmax(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m8))) +vint8m8_t vmax(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m8_m))) +vint8m8_t vmax(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8mf2))) +vint8mf2_t vmax(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8mf2_m))) +vint8mf2_t vmax(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8mf4))) +vint8mf4_t vmax(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8mf4_m))) +vint8mf4_t vmax(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8mf8))) +vint8mf8_t vmax(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8mf8_m))) +vint8mf8_t vmax(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m1))) +vint16m1_t vmax(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m1_m))) +vint16m1_t vmax(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m2))) +vint16m2_t vmax(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m2_m))) +vint16m2_t vmax(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m4))) +vint16m4_t vmax(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m4_m))) +vint16m4_t vmax(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m8))) +vint16m8_t vmax(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m8_m))) +vint16m8_t vmax(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16mf2))) +vint16mf2_t vmax(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16mf2_m))) +vint16mf2_t vmax(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16mf4))) +vint16mf4_t vmax(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16mf4_m))) +vint16mf4_t vmax(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m1))) +vint32m1_t vmax(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m1_m))) +vint32m1_t vmax(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m2))) +vint32m2_t vmax(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m2_m))) +vint32m2_t vmax(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m4))) +vint32m4_t vmax(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m4_m))) +vint32m4_t vmax(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m8))) +vint32m8_t vmax(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m8_m))) +vint32m8_t vmax(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32mf2))) +vint32mf2_t vmax(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32mf2_m))) +vint32mf2_t vmax(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m1))) +vint64m1_t vmax(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m1_m))) +vint64m1_t vmax(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m2))) +vint64m2_t vmax(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m2_m))) +vint64m2_t vmax(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m4))) +vint64m4_t vmax(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m4_m))) +vint64m4_t vmax(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m8))) +vint64m8_t vmax(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m8_m))) +vint64m8_t vmax(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m1))) +vint8m1_t vmul(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m1_m))) +vint8m1_t vmul(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m2))) +vint8m2_t vmul(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m2_m))) +vint8m2_t vmul(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m4))) +vint8m4_t vmul(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m4_m))) +vint8m4_t vmul(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m8))) +vint8m8_t vmul(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m8_m))) +vint8m8_t vmul(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8mf2))) +vint8mf2_t vmul(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8mf2_m))) +vint8mf2_t vmul(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8mf4))) +vint8mf4_t vmul(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8mf4_m))) +vint8mf4_t vmul(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8mf8))) +vint8mf8_t vmul(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8mf8_m))) +vint8mf8_t vmul(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m1))) +vint16m1_t vmul(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m1_m))) +vint16m1_t vmul(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m2))) +vint16m2_t vmul(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m2_m))) +vint16m2_t vmul(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m4))) +vint16m4_t vmul(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m4_m))) +vint16m4_t vmul(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m8))) +vint16m8_t vmul(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m8_m))) +vint16m8_t vmul(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16mf2))) +vint16mf2_t vmul(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16mf2_m))) +vint16mf2_t vmul(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16mf4))) +vint16mf4_t vmul(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16mf4_m))) +vint16mf4_t vmul(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m1))) +vint32m1_t vmul(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m1_m))) +vint32m1_t vmul(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m2))) +vint32m2_t vmul(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m2_m))) +vint32m2_t vmul(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m4))) +vint32m4_t vmul(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m4_m))) +vint32m4_t vmul(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m8))) +vint32m8_t vmul(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m8_m))) +vint32m8_t vmul(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32mf2))) +vint32mf2_t vmul(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32mf2_m))) +vint32mf2_t vmul(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m1))) +vint64m1_t vmul(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m1_m))) +vint64m1_t vmul(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m2))) +vint64m2_t vmul(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m2_m))) +vint64m2_t vmul(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m4))) +vint64m4_t vmul(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m4_m))) +vint64m4_t vmul(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m8))) +vint64m8_t vmul(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m8_m))) +vint64m8_t vmul(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m1))) +vint8m1_t vmul(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m1_m))) +vint8m1_t vmul(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m2))) +vint8m2_t vmul(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m2_m))) +vint8m2_t vmul(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m4))) +vint8m4_t vmul(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m4_m))) +vint8m4_t vmul(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m8))) +vint8m8_t vmul(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m8_m))) +vint8m8_t vmul(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8mf2))) +vint8mf2_t vmul(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8mf2_m))) +vint8mf2_t vmul(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8mf4))) +vint8mf4_t vmul(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8mf4_m))) +vint8mf4_t vmul(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8mf8))) +vint8mf8_t vmul(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8mf8_m))) +vint8mf8_t vmul(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m1))) +vint16m1_t vmul(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m1_m))) +vint16m1_t vmul(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m2))) +vint16m2_t vmul(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m2_m))) +vint16m2_t vmul(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m4))) +vint16m4_t vmul(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m4_m))) +vint16m4_t vmul(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m8))) +vint16m8_t vmul(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m8_m))) +vint16m8_t vmul(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16mf2))) +vint16mf2_t vmul(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16mf2_m))) +vint16mf2_t vmul(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16mf4))) +vint16mf4_t vmul(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16mf4_m))) +vint16mf4_t vmul(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m1))) +vint32m1_t vmul(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m1_m))) +vint32m1_t vmul(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m2))) +vint32m2_t vmul(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m2_m))) +vint32m2_t vmul(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m4))) +vint32m4_t vmul(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m4_m))) +vint32m4_t vmul(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m8))) +vint32m8_t vmul(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m8_m))) +vint32m8_t vmul(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32mf2))) +vint32mf2_t vmul(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32mf2_m))) +vint32mf2_t vmul(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m1))) +vint64m1_t vmul(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m1_m))) +vint64m1_t vmul(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m2))) +vint64m2_t vmul(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m2_m))) +vint64m2_t vmul(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m4))) +vint64m4_t vmul(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m4_m))) +vint64m4_t vmul(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m8))) +vint64m8_t vmul(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m8_m))) +vint64m8_t vmul(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m1))) +vuint8m1_t vmul(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m1_m))) +vuint8m1_t vmul(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m2))) +vuint8m2_t vmul(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m2_m))) +vuint8m2_t vmul(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m4))) +vuint8m4_t vmul(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m4_m))) +vuint8m4_t vmul(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m8))) +vuint8m8_t vmul(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m8_m))) +vuint8m8_t vmul(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8mf2))) +vuint8mf2_t vmul(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8mf2_m))) +vuint8mf2_t vmul(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8mf4))) +vuint8mf4_t vmul(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8mf4_m))) +vuint8mf4_t vmul(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8mf8))) +vuint8mf8_t vmul(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8mf8_m))) +vuint8mf8_t vmul(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m1))) +vuint16m1_t vmul(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m1_m))) +vuint16m1_t vmul(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m2))) +vuint16m2_t vmul(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m2_m))) +vuint16m2_t vmul(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m4))) +vuint16m4_t vmul(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m4_m))) +vuint16m4_t vmul(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m8))) +vuint16m8_t vmul(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m8_m))) +vuint16m8_t vmul(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16mf2))) +vuint16mf2_t vmul(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16mf2_m))) +vuint16mf2_t vmul(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16mf4))) +vuint16mf4_t vmul(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16mf4_m))) +vuint16mf4_t vmul(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m1))) +vuint32m1_t vmul(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m1_m))) +vuint32m1_t vmul(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m2))) +vuint32m2_t vmul(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m2_m))) +vuint32m2_t vmul(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m4))) +vuint32m4_t vmul(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m4_m))) +vuint32m4_t vmul(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m8))) +vuint32m8_t vmul(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m8_m))) +vuint32m8_t vmul(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32mf2))) +vuint32mf2_t vmul(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32mf2_m))) +vuint32mf2_t vmul(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m1))) +vuint64m1_t vmul(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m1_m))) +vuint64m1_t vmul(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m2))) +vuint64m2_t vmul(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m2_m))) +vuint64m2_t vmul(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m4))) +vuint64m4_t vmul(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m4_m))) +vuint64m4_t vmul(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m8))) +vuint64m8_t vmul(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m8_m))) +vuint64m8_t vmul(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m1))) +vuint8m1_t vmul(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m1_m))) +vuint8m1_t vmul(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m2))) +vuint8m2_t vmul(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m2_m))) +vuint8m2_t vmul(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m4))) +vuint8m4_t vmul(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m4_m))) +vuint8m4_t vmul(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m8))) +vuint8m8_t vmul(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m8_m))) +vuint8m8_t vmul(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8mf2))) +vuint8mf2_t vmul(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8mf2_m))) +vuint8mf2_t vmul(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8mf4))) +vuint8mf4_t vmul(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8mf4_m))) +vuint8mf4_t vmul(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8mf8))) +vuint8mf8_t vmul(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8mf8_m))) +vuint8mf8_t vmul(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m1))) +vuint16m1_t vmul(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m1_m))) +vuint16m1_t vmul(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m2))) +vuint16m2_t vmul(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m2_m))) +vuint16m2_t vmul(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m4))) +vuint16m4_t vmul(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m4_m))) +vuint16m4_t vmul(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m8))) +vuint16m8_t vmul(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m8_m))) +vuint16m8_t vmul(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16mf2))) +vuint16mf2_t vmul(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16mf2_m))) +vuint16mf2_t vmul(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16mf4))) +vuint16mf4_t vmul(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16mf4_m))) +vuint16mf4_t vmul(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m1))) +vuint32m1_t vmul(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m1_m))) +vuint32m1_t vmul(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m2))) +vuint32m2_t vmul(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m2_m))) +vuint32m2_t vmul(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m4))) +vuint32m4_t vmul(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m4_m))) +vuint32m4_t vmul(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m8))) +vuint32m8_t vmul(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m8_m))) +vuint32m8_t vmul(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32mf2))) +vuint32mf2_t vmul(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32mf2_m))) +vuint32mf2_t vmul(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m1))) +vuint64m1_t vmul(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m1_m))) +vuint64m1_t vmul(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m2))) +vuint64m2_t vmul(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m2_m))) +vuint64m2_t vmul(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m4))) +vuint64m4_t vmul(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m4_m))) +vuint64m4_t vmul(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m8))) +vuint64m8_t vmul(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m8_m))) +vuint64m8_t vmul(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m1))) +vint8m1_t vmulh(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m1_m))) +vint8m1_t vmulh(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m2))) +vint8m2_t vmulh(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m2_m))) +vint8m2_t vmulh(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m4))) +vint8m4_t vmulh(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m4_m))) +vint8m4_t vmulh(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m8))) +vint8m8_t vmulh(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m8_m))) +vint8m8_t vmulh(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8mf2))) +vint8mf2_t vmulh(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8mf2_m))) +vint8mf2_t vmulh(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8mf4))) +vint8mf4_t vmulh(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8mf4_m))) +vint8mf4_t vmulh(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8mf8))) +vint8mf8_t vmulh(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8mf8_m))) +vint8mf8_t vmulh(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m1))) +vint16m1_t vmulh(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m1_m))) +vint16m1_t vmulh(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m2))) +vint16m2_t vmulh(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m2_m))) +vint16m2_t vmulh(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m4))) +vint16m4_t vmulh(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m4_m))) +vint16m4_t vmulh(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m8))) +vint16m8_t vmulh(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m8_m))) +vint16m8_t vmulh(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16mf2))) +vint16mf2_t vmulh(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16mf2_m))) +vint16mf2_t vmulh(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16mf4))) +vint16mf4_t vmulh(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16mf4_m))) +vint16mf4_t vmulh(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m1))) +vint32m1_t vmulh(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m1_m))) +vint32m1_t vmulh(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m2))) +vint32m2_t vmulh(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m2_m))) +vint32m2_t vmulh(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m4))) +vint32m4_t vmulh(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m4_m))) +vint32m4_t vmulh(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m8))) +vint32m8_t vmulh(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m8_m))) +vint32m8_t vmulh(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32mf2))) +vint32mf2_t vmulh(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32mf2_m))) +vint32mf2_t vmulh(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m1))) +vint64m1_t vmulh(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m1_m))) +vint64m1_t vmulh(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m2))) +vint64m2_t vmulh(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m2_m))) +vint64m2_t vmulh(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m4))) +vint64m4_t vmulh(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m4_m))) +vint64m4_t vmulh(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m8))) +vint64m8_t vmulh(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m8_m))) +vint64m8_t vmulh(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m1))) +vint8m1_t vmulh(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m1_m))) +vint8m1_t vmulh(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m2))) +vint8m2_t vmulh(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m2_m))) +vint8m2_t vmulh(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m4))) +vint8m4_t vmulh(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m4_m))) +vint8m4_t vmulh(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m8))) +vint8m8_t vmulh(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m8_m))) +vint8m8_t vmulh(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8mf2))) +vint8mf2_t vmulh(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8mf2_m))) +vint8mf2_t vmulh(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8mf4))) +vint8mf4_t vmulh(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8mf4_m))) +vint8mf4_t vmulh(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8mf8))) +vint8mf8_t vmulh(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8mf8_m))) +vint8mf8_t vmulh(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m1))) +vint16m1_t vmulh(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m1_m))) +vint16m1_t vmulh(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m2))) +vint16m2_t vmulh(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m2_m))) +vint16m2_t vmulh(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m4))) +vint16m4_t vmulh(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m4_m))) +vint16m4_t vmulh(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m8))) +vint16m8_t vmulh(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m8_m))) +vint16m8_t vmulh(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16mf2))) +vint16mf2_t vmulh(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16mf2_m))) +vint16mf2_t vmulh(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16mf4))) +vint16mf4_t vmulh(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16mf4_m))) +vint16mf4_t vmulh(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m1))) +vint32m1_t vmulh(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m1_m))) +vint32m1_t vmulh(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m2))) +vint32m2_t vmulh(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m2_m))) +vint32m2_t vmulh(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m4))) +vint32m4_t vmulh(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m4_m))) +vint32m4_t vmulh(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m8))) +vint32m8_t vmulh(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m8_m))) +vint32m8_t vmulh(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32mf2))) +vint32mf2_t vmulh(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32mf2_m))) +vint32mf2_t vmulh(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m1))) +vint64m1_t vmulh(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m1_m))) +vint64m1_t vmulh(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m2))) +vint64m2_t vmulh(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m2_m))) +vint64m2_t vmulh(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m4))) +vint64m4_t vmulh(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m4_m))) +vint64m4_t vmulh(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m8))) +vint64m8_t vmulh(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m8_m))) +vint64m8_t vmulh(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m1))) +vuint8m1_t vmulhu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m1_m))) +vuint8m1_t vmulhu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m2))) +vuint8m2_t vmulhu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m2_m))) +vuint8m2_t vmulhu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m4))) +vuint8m4_t vmulhu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m4_m))) +vuint8m4_t vmulhu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m8))) +vuint8m8_t vmulhu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m8_m))) +vuint8m8_t vmulhu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8mf2))) +vuint8mf2_t vmulhu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8mf2_m))) +vuint8mf2_t vmulhu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8mf4))) +vuint8mf4_t vmulhu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8mf4_m))) +vuint8mf4_t vmulhu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8mf8))) +vuint8mf8_t vmulhu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8mf8_m))) +vuint8mf8_t vmulhu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m1))) +vuint16m1_t vmulhu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m1_m))) +vuint16m1_t vmulhu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m2))) +vuint16m2_t vmulhu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m2_m))) +vuint16m2_t vmulhu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m4))) +vuint16m4_t vmulhu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m4_m))) +vuint16m4_t vmulhu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m8))) +vuint16m8_t vmulhu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m8_m))) +vuint16m8_t vmulhu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16mf2))) +vuint16mf2_t vmulhu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16mf2_m))) +vuint16mf2_t vmulhu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16mf4))) +vuint16mf4_t vmulhu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16mf4_m))) +vuint16mf4_t vmulhu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m1))) +vuint32m1_t vmulhu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m1_m))) +vuint32m1_t vmulhu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m2))) +vuint32m2_t vmulhu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m2_m))) +vuint32m2_t vmulhu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m4))) +vuint32m4_t vmulhu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m4_m))) +vuint32m4_t vmulhu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m8))) +vuint32m8_t vmulhu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m8_m))) +vuint32m8_t vmulhu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32mf2))) +vuint32mf2_t vmulhu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32mf2_m))) +vuint32mf2_t vmulhu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m1))) +vuint64m1_t vmulhu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m1_m))) +vuint64m1_t vmulhu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m2))) +vuint64m2_t vmulhu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m2_m))) +vuint64m2_t vmulhu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m4))) +vuint64m4_t vmulhu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m4_m))) +vuint64m4_t vmulhu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m8))) +vuint64m8_t vmulhu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m8_m))) +vuint64m8_t vmulhu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m1))) +vuint8m1_t vmulhu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m1_m))) +vuint8m1_t vmulhu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m2))) +vuint8m2_t vmulhu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m2_m))) +vuint8m2_t vmulhu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m4))) +vuint8m4_t vmulhu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m4_m))) +vuint8m4_t vmulhu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m8))) +vuint8m8_t vmulhu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m8_m))) +vuint8m8_t vmulhu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8mf2))) +vuint8mf2_t vmulhu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8mf2_m))) +vuint8mf2_t vmulhu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8mf4))) +vuint8mf4_t vmulhu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8mf4_m))) +vuint8mf4_t vmulhu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8mf8))) +vuint8mf8_t vmulhu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8mf8_m))) +vuint8mf8_t vmulhu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m1))) +vuint16m1_t vmulhu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m1_m))) +vuint16m1_t vmulhu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m2))) +vuint16m2_t vmulhu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m2_m))) +vuint16m2_t vmulhu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m4))) +vuint16m4_t vmulhu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m4_m))) +vuint16m4_t vmulhu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m8))) +vuint16m8_t vmulhu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m8_m))) +vuint16m8_t vmulhu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16mf2))) +vuint16mf2_t vmulhu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16mf2_m))) +vuint16mf2_t vmulhu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16mf4))) +vuint16mf4_t vmulhu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16mf4_m))) +vuint16mf4_t vmulhu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m1))) +vuint32m1_t vmulhu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m1_m))) +vuint32m1_t vmulhu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m2))) +vuint32m2_t vmulhu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m2_m))) +vuint32m2_t vmulhu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m4))) +vuint32m4_t vmulhu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m4_m))) +vuint32m4_t vmulhu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m8))) +vuint32m8_t vmulhu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m8_m))) +vuint32m8_t vmulhu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32mf2))) +vuint32mf2_t vmulhu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32mf2_m))) +vuint32mf2_t vmulhu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m1))) +vuint64m1_t vmulhu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m1_m))) +vuint64m1_t vmulhu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m2))) +vuint64m2_t vmulhu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m2_m))) +vuint64m2_t vmulhu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m4))) +vuint64m4_t vmulhu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m4_m))) +vuint64m4_t vmulhu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m8))) +vuint64m8_t vmulhu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m8_m))) +vuint64m8_t vmulhu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m1))) +vint8m1_t vmulhsu(vint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m1_m))) +vint8m1_t vmulhsu(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m2))) +vint8m2_t vmulhsu(vint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m2_m))) +vint8m2_t vmulhsu(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m4))) +vint8m4_t vmulhsu(vint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m4_m))) +vint8m4_t vmulhsu(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m8))) +vint8m8_t vmulhsu(vint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m8_m))) +vint8m8_t vmulhsu(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8mf2))) +vint8mf2_t vmulhsu(vint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8mf2_m))) +vint8mf2_t vmulhsu(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8mf4))) +vint8mf4_t vmulhsu(vint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8mf4_m))) +vint8mf4_t vmulhsu(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8mf8))) +vint8mf8_t vmulhsu(vint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8mf8_m))) +vint8mf8_t vmulhsu(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m1))) +vint16m1_t vmulhsu(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m1_m))) +vint16m1_t vmulhsu(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m2))) +vint16m2_t vmulhsu(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m2_m))) +vint16m2_t vmulhsu(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m4))) +vint16m4_t vmulhsu(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m4_m))) +vint16m4_t vmulhsu(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m8))) +vint16m8_t vmulhsu(vint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m8_m))) +vint16m8_t vmulhsu(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16mf2))) +vint16mf2_t vmulhsu(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16mf2_m))) +vint16mf2_t vmulhsu(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16mf4))) +vint16mf4_t vmulhsu(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16mf4_m))) +vint16mf4_t vmulhsu(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m1))) +vint32m1_t vmulhsu(vint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m1_m))) +vint32m1_t vmulhsu(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m2))) +vint32m2_t vmulhsu(vint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m2_m))) +vint32m2_t vmulhsu(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m4))) +vint32m4_t vmulhsu(vint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m4_m))) +vint32m4_t vmulhsu(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m8))) +vint32m8_t vmulhsu(vint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m8_m))) +vint32m8_t vmulhsu(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32mf2))) +vint32mf2_t vmulhsu(vint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32mf2_m))) +vint32mf2_t vmulhsu(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m1))) +vint64m1_t vmulhsu(vint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m1_m))) +vint64m1_t vmulhsu(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m2))) +vint64m2_t vmulhsu(vint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m2_m))) +vint64m2_t vmulhsu(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m4))) +vint64m4_t vmulhsu(vint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m4_m))) +vint64m4_t vmulhsu(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m8))) +vint64m8_t vmulhsu(vint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m8_m))) +vint64m8_t vmulhsu(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m1))) +void vsuxei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m1_m))) +void vsuxei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m2))) +void vsuxei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m2_m))) +void vsuxei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m4))) +void vsuxei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m4_m))) +void vsuxei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m8))) +void vsuxei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m8_m))) +void vsuxei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32mf2))) +void vsuxei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32mf2_m))) +void vsuxei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m1))) +vint8m1_t vmulhsu(vint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m1_m))) +vint8m1_t vmulhsu(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m2))) +vint8m2_t vmulhsu(vint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m2_m))) +vint8m2_t vmulhsu(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m4))) +vint8m4_t vmulhsu(vint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m4_m))) +vint8m4_t vmulhsu(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m8))) +vint8m8_t vmulhsu(vint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m8_m))) +vint8m8_t vmulhsu(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8mf2))) +vint8mf2_t vmulhsu(vint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8mf2_m))) +vint8mf2_t vmulhsu(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8mf4))) +vint8mf4_t vmulhsu(vint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8mf4_m))) +vint8mf4_t vmulhsu(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8mf8))) +vint8mf8_t vmulhsu(vint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8mf8_m))) +vint8mf8_t vmulhsu(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m1))) +vint16m1_t vmulhsu(vint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m1_m))) +vint16m1_t vmulhsu(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m2))) +vint16m2_t vmulhsu(vint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m2_m))) +vint16m2_t vmulhsu(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m4))) +vint16m4_t vmulhsu(vint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m4_m))) +vint16m4_t vmulhsu(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m8))) +vint16m8_t vmulhsu(vint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m8_m))) +vint16m8_t vmulhsu(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16mf2))) +vint16mf2_t vmulhsu(vint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16mf2_m))) +vint16mf2_t vmulhsu(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16mf4))) +vint16mf4_t vmulhsu(vint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16mf4_m))) +vint16mf4_t vmulhsu(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m1))) +vint32m1_t vmulhsu(vint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m1_m))) +vint32m1_t vmulhsu(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m2))) +vint32m2_t vmulhsu(vint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m2_m))) +vint32m2_t vmulhsu(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m4))) +vint32m4_t vmulhsu(vint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m4_m))) +vint32m4_t vmulhsu(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m8))) +vint32m8_t vmulhsu(vint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m8_m))) +vint32m8_t vmulhsu(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32mf2))) +vint32mf2_t vmulhsu(vint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32mf2_m))) +vint32mf2_t vmulhsu(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m1))) +vint64m1_t vmulhsu(vint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m1_m))) +vint64m1_t vmulhsu(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m2))) +vint64m2_t vmulhsu(vint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m2_m))) +vint64m2_t vmulhsu(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m4))) +vint64m4_t vmulhsu(vint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m4_m))) +vint64m4_t vmulhsu(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m8))) +vint64m8_t vmulhsu(vint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m8_m))) +vint64m8_t vmulhsu(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m1))) +vuint8m1_t vdivu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m1_m))) +vuint8m1_t vdivu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m2))) +vuint8m2_t vdivu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m2_m))) +vuint8m2_t vdivu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m4))) +vuint8m4_t vdivu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m4_m))) +vuint8m4_t vdivu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m8))) +vuint8m8_t vdivu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m8_m))) +vuint8m8_t vdivu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8mf2))) +vuint8mf2_t vdivu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8mf2_m))) +vuint8mf2_t vdivu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8mf4))) +vuint8mf4_t vdivu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8mf4_m))) +vuint8mf4_t vdivu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8mf8))) +vuint8mf8_t vdivu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8mf8_m))) +vuint8mf8_t vdivu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m1))) +vuint16m1_t vdivu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m1_m))) +vuint16m1_t vdivu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m2))) +vuint16m2_t vdivu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m2_m))) +vuint16m2_t vdivu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m4))) +vuint16m4_t vdivu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m4_m))) +vuint16m4_t vdivu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m8))) +vuint16m8_t vdivu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m8_m))) +vuint16m8_t vdivu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16mf2))) +vuint16mf2_t vdivu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16mf2_m))) +vuint16mf2_t vdivu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16mf4))) +vuint16mf4_t vdivu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16mf4_m))) +vuint16mf4_t vdivu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m1))) +vuint32m1_t vdivu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m1_m))) +vuint32m1_t vdivu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m2))) +vuint32m2_t vdivu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m2_m))) +vuint32m2_t vdivu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m4))) +vuint32m4_t vdivu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m4_m))) +vuint32m4_t vdivu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m8))) +vuint32m8_t vdivu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m8_m))) +vuint32m8_t vdivu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32mf2))) +vuint32mf2_t vdivu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32mf2_m))) +vuint32mf2_t vdivu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m1))) +vuint64m1_t vdivu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m1_m))) +vuint64m1_t vdivu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m2))) +vuint64m2_t vdivu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m2_m))) +vuint64m2_t vdivu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m4))) +vuint64m4_t vdivu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m4_m))) +vuint64m4_t vdivu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m8))) +vuint64m8_t vdivu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m8_m))) +vuint64m8_t vdivu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m1))) +vuint8m1_t vdivu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m1_m))) +vuint8m1_t vdivu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m2))) +vuint8m2_t vdivu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m2_m))) +vuint8m2_t vdivu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m4))) +vuint8m4_t vdivu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m4_m))) +vuint8m4_t vdivu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m8))) +vuint8m8_t vdivu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m8_m))) +vuint8m8_t vdivu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8mf2))) +vuint8mf2_t vdivu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8mf2_m))) +vuint8mf2_t vdivu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8mf4))) +vuint8mf4_t vdivu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8mf4_m))) +vuint8mf4_t vdivu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8mf8))) +vuint8mf8_t vdivu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8mf8_m))) +vuint8mf8_t vdivu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m1))) +vuint16m1_t vdivu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m1_m))) +vuint16m1_t vdivu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m2))) +vuint16m2_t vdivu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m2_m))) +vuint16m2_t vdivu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m4))) +vuint16m4_t vdivu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m4_m))) +vuint16m4_t vdivu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m8))) +vuint16m8_t vdivu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m8_m))) +vuint16m8_t vdivu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16mf2))) +vuint16mf2_t vdivu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16mf2_m))) +vuint16mf2_t vdivu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16mf4))) +vuint16mf4_t vdivu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16mf4_m))) +vuint16mf4_t vdivu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m1))) +vuint32m1_t vdivu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m1_m))) +vuint32m1_t vdivu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m2))) +vuint32m2_t vdivu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m2_m))) +vuint32m2_t vdivu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m4))) +vuint32m4_t vdivu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m4_m))) +vuint32m4_t vdivu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m8))) +vuint32m8_t vdivu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m8_m))) +vuint32m8_t vdivu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32mf2))) +vuint32mf2_t vdivu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32mf2_m))) +vuint32mf2_t vdivu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m1))) +vuint64m1_t vdivu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m1_m))) +vuint64m1_t vdivu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m2))) +vuint64m2_t vdivu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m2_m))) +vuint64m2_t vdivu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m4))) +vuint64m4_t vdivu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m4_m))) +vuint64m4_t vdivu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m8))) +vuint64m8_t vdivu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m8_m))) +vuint64m8_t vdivu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m1))) +vint8m1_t vdiv(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m1_m))) +vint8m1_t vdiv(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m2))) +vint8m2_t vdiv(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m2_m))) +vint8m2_t vdiv(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m4))) +vint8m4_t vdiv(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m4_m))) +vint8m4_t vdiv(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m8))) +vint8m8_t vdiv(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m8_m))) +vint8m8_t vdiv(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8mf2))) +vint8mf2_t vdiv(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8mf2_m))) +vint8mf2_t vdiv(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8mf4))) +vint8mf4_t vdiv(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8mf4_m))) +vint8mf4_t vdiv(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8mf8))) +vint8mf8_t vdiv(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8mf8_m))) +vint8mf8_t vdiv(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m1))) +vint16m1_t vdiv(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m1_m))) +vint16m1_t vdiv(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m2))) +vint16m2_t vdiv(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m2_m))) +vint16m2_t vdiv(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m4))) +vint16m4_t vdiv(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m4_m))) +vint16m4_t vdiv(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m8))) +vint16m8_t vdiv(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m8_m))) +vint16m8_t vdiv(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16mf2))) +vint16mf2_t vdiv(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16mf2_m))) +vint16mf2_t vdiv(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16mf4))) +vint16mf4_t vdiv(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16mf4_m))) +vint16mf4_t vdiv(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m1))) +vint32m1_t vdiv(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m1_m))) +vint32m1_t vdiv(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m2))) +vint32m2_t vdiv(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m2_m))) +vint32m2_t vdiv(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m4))) +vint32m4_t vdiv(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m4_m))) +vint32m4_t vdiv(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m8))) +vint32m8_t vdiv(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m8_m))) +vint32m8_t vdiv(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32mf2))) +vint32mf2_t vdiv(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32mf2_m))) +vint32mf2_t vdiv(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m1))) +vint64m1_t vdiv(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m1_m))) +vint64m1_t vdiv(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m2))) +vint64m2_t vdiv(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m2_m))) +vint64m2_t vdiv(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m4))) +vint64m4_t vdiv(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m4_m))) +vint64m4_t vdiv(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m8))) +vint64m8_t vdiv(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m8_m))) +vint64m8_t vdiv(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m1))) +vint8m1_t vdiv(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m1_m))) +vint8m1_t vdiv(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m2))) +vint8m2_t vdiv(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m2_m))) +vint8m2_t vdiv(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m4))) +vint8m4_t vdiv(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m4_m))) +vint8m4_t vdiv(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m8))) +vint8m8_t vdiv(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m8_m))) +vint8m8_t vdiv(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8mf2))) +vint8mf2_t vdiv(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8mf2_m))) +vint8mf2_t vdiv(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8mf4))) +vint8mf4_t vdiv(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8mf4_m))) +vint8mf4_t vdiv(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8mf8))) +vint8mf8_t vdiv(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8mf8_m))) +vint8mf8_t vdiv(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m1))) +vint16m1_t vdiv(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m1_m))) +vint16m1_t vdiv(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m2))) +vint16m2_t vdiv(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m2_m))) +vint16m2_t vdiv(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m4))) +vint16m4_t vdiv(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m4_m))) +vint16m4_t vdiv(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m8))) +vint16m8_t vdiv(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m8_m))) +vint16m8_t vdiv(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16mf2))) +vint16mf2_t vdiv(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16mf2_m))) +vint16mf2_t vdiv(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16mf4))) +vint16mf4_t vdiv(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16mf4_m))) +vint16mf4_t vdiv(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m1))) +vint32m1_t vdiv(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m1_m))) +vint32m1_t vdiv(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m2))) +vint32m2_t vdiv(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m2_m))) +vint32m2_t vdiv(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m4))) +vint32m4_t vdiv(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m4_m))) +vint32m4_t vdiv(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m8))) +vint32m8_t vdiv(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m8_m))) +vint32m8_t vdiv(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32mf2))) +vint32mf2_t vdiv(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32mf2_m))) +vint32mf2_t vdiv(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m1))) +vint64m1_t vdiv(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m1_m))) +vint64m1_t vdiv(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m2))) +vint64m2_t vdiv(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m2_m))) +vint64m2_t vdiv(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m4))) +vint64m4_t vdiv(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m4_m))) +vint64m4_t vdiv(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m8))) +vint64m8_t vdiv(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m8_m))) +vint64m8_t vdiv(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m1))) +vuint8m1_t vremu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m1_m))) +vuint8m1_t vremu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m2))) +vuint8m2_t vremu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m2_m))) +vuint8m2_t vremu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m4))) +vuint8m4_t vremu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m4_m))) +vuint8m4_t vremu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m8))) +vuint8m8_t vremu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m8_m))) +vuint8m8_t vremu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8mf2))) +vuint8mf2_t vremu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8mf2_m))) +vuint8mf2_t vremu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8mf4))) +vuint8mf4_t vremu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8mf4_m))) +vuint8mf4_t vremu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8mf8))) +vuint8mf8_t vremu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8mf8_m))) +vuint8mf8_t vremu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m1))) +vuint16m1_t vremu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m1_m))) +vuint16m1_t vremu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m2))) +vuint16m2_t vremu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m2_m))) +vuint16m2_t vremu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m4))) +vuint16m4_t vremu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m4_m))) +vuint16m4_t vremu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m8))) +vuint16m8_t vremu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m8_m))) +vuint16m8_t vremu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16mf2))) +vuint16mf2_t vremu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16mf2_m))) +vuint16mf2_t vremu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16mf4))) +vuint16mf4_t vremu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16mf4_m))) +vuint16mf4_t vremu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m1))) +vuint32m1_t vremu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m1_m))) +vuint32m1_t vremu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m2))) +vuint32m2_t vremu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m2_m))) +vuint32m2_t vremu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m4))) +vuint32m4_t vremu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m4_m))) +vuint32m4_t vremu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m8))) +vuint32m8_t vremu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m8_m))) +vuint32m8_t vremu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32mf2))) +vuint32mf2_t vremu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32mf2_m))) +vuint32mf2_t vremu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m1))) +vuint64m1_t vremu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m1_m))) +vuint64m1_t vremu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m2))) +vuint64m2_t vremu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m2_m))) +vuint64m2_t vremu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m4))) +vuint64m4_t vremu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m4_m))) +vuint64m4_t vremu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m8))) +vuint64m8_t vremu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m8_m))) +vuint64m8_t vremu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m1))) +vuint8m1_t vremu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m1_m))) +vuint8m1_t vremu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m2))) +vuint8m2_t vremu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m2_m))) +vuint8m2_t vremu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m4))) +vuint8m4_t vremu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m4_m))) +vuint8m4_t vremu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m8))) +vuint8m8_t vremu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m8_m))) +vuint8m8_t vremu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8mf2))) +vuint8mf2_t vremu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8mf2_m))) +vuint8mf2_t vremu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8mf4))) +vuint8mf4_t vremu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8mf4_m))) +vuint8mf4_t vremu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8mf8))) +vuint8mf8_t vremu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8mf8_m))) +vuint8mf8_t vremu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m1))) +vuint16m1_t vremu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m1_m))) +vuint16m1_t vremu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m2))) +vuint16m2_t vremu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m2_m))) +vuint16m2_t vremu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m4))) +vuint16m4_t vremu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m4_m))) +vuint16m4_t vremu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m8))) +vuint16m8_t vremu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m8_m))) +vuint16m8_t vremu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16mf2))) +vuint16mf2_t vremu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16mf2_m))) +vuint16mf2_t vremu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16mf4))) +vuint16mf4_t vremu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16mf4_m))) +vuint16mf4_t vremu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m1))) +vuint32m1_t vremu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m1_m))) +vuint32m1_t vremu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m2))) +vuint32m2_t vremu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m2_m))) +vuint32m2_t vremu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m4))) +vuint32m4_t vremu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m4_m))) +vuint32m4_t vremu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m8))) +vuint32m8_t vremu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m8_m))) +vuint32m8_t vremu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32mf2))) +vuint32mf2_t vremu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32mf2_m))) +vuint32mf2_t vremu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m1))) +vuint64m1_t vremu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m1_m))) +vuint64m1_t vremu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m2))) +vuint64m2_t vremu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m2_m))) +vuint64m2_t vremu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m4))) +vuint64m4_t vremu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m4_m))) +vuint64m4_t vremu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m8))) +vuint64m8_t vremu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m8_m))) +vuint64m8_t vremu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m1))) +vint8m1_t vrem(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m1_m))) +vint8m1_t vrem(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m2))) +vint8m2_t vrem(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m2_m))) +vint8m2_t vrem(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m4))) +vint8m4_t vrem(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m4_m))) +vint8m4_t vrem(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m8))) +vint8m8_t vrem(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m8_m))) +vint8m8_t vrem(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8mf2))) +vint8mf2_t vrem(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8mf2_m))) +vint8mf2_t vrem(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8mf4))) +vint8mf4_t vrem(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8mf4_m))) +vint8mf4_t vrem(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8mf8))) +vint8mf8_t vrem(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8mf8_m))) +vint8mf8_t vrem(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m1))) +vint16m1_t vrem(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m1_m))) +vint16m1_t vrem(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m2))) +vint16m2_t vrem(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m2_m))) +vint16m2_t vrem(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m4))) +vint16m4_t vrem(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m4_m))) +vint16m4_t vrem(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m8))) +vint16m8_t vrem(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m8_m))) +vint16m8_t vrem(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16mf2))) +vint16mf2_t vrem(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16mf2_m))) +vint16mf2_t vrem(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16mf4))) +vint16mf4_t vrem(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16mf4_m))) +vint16mf4_t vrem(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m1))) +vint32m1_t vrem(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m1_m))) +vint32m1_t vrem(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m2))) +vint32m2_t vrem(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m2_m))) +vint32m2_t vrem(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m4))) +vint32m4_t vrem(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m4_m))) +vint32m4_t vrem(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m8))) +vint32m8_t vrem(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m8_m))) +vint32m8_t vrem(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32mf2))) +vint32mf2_t vrem(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32mf2_m))) +vint32mf2_t vrem(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m1))) +vint64m1_t vrem(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m1_m))) +vint64m1_t vrem(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m2))) +vint64m2_t vrem(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m2_m))) +vint64m2_t vrem(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m4))) +vint64m4_t vrem(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m4_m))) +vint64m4_t vrem(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m8))) +vint64m8_t vrem(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m8_m))) +vint64m8_t vrem(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m1))) +vint8m1_t vrem(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m1_m))) +vint8m1_t vrem(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m2))) +vint8m2_t vrem(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m2_m))) +vint8m2_t vrem(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m4))) +vint8m4_t vrem(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m4_m))) +vint8m4_t vrem(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m8))) +vint8m8_t vrem(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m8_m))) +vint8m8_t vrem(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8mf2))) +vint8mf2_t vrem(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8mf2_m))) +vint8mf2_t vrem(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8mf4))) +vint8mf4_t vrem(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8mf4_m))) +vint8mf4_t vrem(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8mf8))) +vint8mf8_t vrem(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8mf8_m))) +vint8mf8_t vrem(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m1))) +vint16m1_t vrem(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m1_m))) +vint16m1_t vrem(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m2))) +vint16m2_t vrem(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m2_m))) +vint16m2_t vrem(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m4))) +vint16m4_t vrem(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m4_m))) +vint16m4_t vrem(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m8))) +vint16m8_t vrem(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m8_m))) +vint16m8_t vrem(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16mf2))) +vint16mf2_t vrem(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16mf2_m))) +vint16mf2_t vrem(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16mf4))) +vint16mf4_t vrem(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16mf4_m))) +vint16mf4_t vrem(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m1))) +vint32m1_t vrem(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m1_m))) +vint32m1_t vrem(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m2))) +vint32m2_t vrem(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m2_m))) +vint32m2_t vrem(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m4))) +vint32m4_t vrem(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m4_m))) +vint32m4_t vrem(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m8))) +vint32m8_t vrem(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m8_m))) +vint32m8_t vrem(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32mf2))) +vint32mf2_t vrem(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32mf2_m))) +vint32mf2_t vrem(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m1))) +vint64m1_t vrem(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m1_m))) +vint64m1_t vrem(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m2))) +vint64m2_t vrem(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m2_m))) +vint64m2_t vrem(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m4))) +vint64m4_t vrem(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m4_m))) +vint64m4_t vrem(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m8))) +vint64m8_t vrem(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m8_m))) +vint64m8_t vrem(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16mf4))) +vint16mf4_t vwmul(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16mf4_m))) +vint16mf4_t vwmul(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16mf2))) +vint16mf2_t vwmul(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16mf2_m))) +vint16mf2_t vwmul(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m1))) +vint16m1_t vwmul(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m1_m))) +vint16m1_t vwmul(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m2))) +vint16m2_t vwmul(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m2_m))) +vint16m2_t vwmul(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m4))) +vint16m4_t vwmul(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m4_m))) +vint16m4_t vwmul(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m8))) +vint16m8_t vwmul(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m8_m))) +vint16m8_t vwmul(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32mf2))) +vint32mf2_t vwmul(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32mf2_m))) +vint32mf2_t vwmul(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m1))) +vint32m1_t vwmul(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m1_m))) +vint32m1_t vwmul(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m2))) +vint32m2_t vwmul(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m2_m))) +vint32m2_t vwmul(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m4))) +vint32m4_t vwmul(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m4_m))) +vint32m4_t vwmul(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m8))) +vint32m8_t vwmul(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m8_m))) +vint32m8_t vwmul(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m1))) +vint64m1_t vwmul(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m1_m))) +vint64m1_t vwmul(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m2))) +vint64m2_t vwmul(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m2_m))) +vint64m2_t vwmul(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m4))) +vint64m4_t vwmul(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m4_m))) +vint64m4_t vwmul(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m8))) +vint64m8_t vwmul(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m8_m))) +vint64m8_t vwmul(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m1))) +void vsuxei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m1_m))) +void vsuxei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m2))) +void vsuxei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m2_m))) +void vsuxei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m4))) +void vsuxei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m4_m))) +void vsuxei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m8))) +void vsuxei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m8_m))) +void vsuxei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32mf2))) +void vsuxei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32mf2_m))) +void vsuxei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16mf4))) +vint16mf4_t vwmul(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16mf4_m))) +vint16mf4_t vwmul(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16mf2))) +vint16mf2_t vwmul(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16mf2_m))) +vint16mf2_t vwmul(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m1))) +vint16m1_t vwmul(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m1_m))) +vint16m1_t vwmul(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m2))) +vint16m2_t vwmul(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m2_m))) +vint16m2_t vwmul(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m4))) +vint16m4_t vwmul(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m4_m))) +vint16m4_t vwmul(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m8))) +vint16m8_t vwmul(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m8_m))) +vint16m8_t vwmul(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32mf2))) +vint32mf2_t vwmul(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32mf2_m))) +vint32mf2_t vwmul(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m1))) +vint32m1_t vwmul(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m1_m))) +vint32m1_t vwmul(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m2))) +vint32m2_t vwmul(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m2_m))) +vint32m2_t vwmul(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m4))) +vint32m4_t vwmul(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m4_m))) +vint32m4_t vwmul(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m8))) +vint32m8_t vwmul(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m8_m))) +vint32m8_t vwmul(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m1))) +vint64m1_t vwmul(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m1_m))) +vint64m1_t vwmul(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m2))) +vint64m2_t vwmul(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m2_m))) +vint64m2_t vwmul(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m4))) +vint64m4_t vwmul(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m4_m))) +vint64m4_t vwmul(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m8))) +vint64m8_t vwmul(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m8_m))) +vint64m8_t vwmul(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16mf4))) +vuint16mf4_t vwmulu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16mf4_m))) +vuint16mf4_t vwmulu(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16mf2))) +vuint16mf2_t vwmulu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16mf2_m))) +vuint16mf2_t vwmulu(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m1))) +vuint16m1_t vwmulu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m1_m))) +vuint16m1_t vwmulu(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m2))) +vuint16m2_t vwmulu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m2_m))) +vuint16m2_t vwmulu(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m4))) +vuint16m4_t vwmulu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m4_m))) +vuint16m4_t vwmulu(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m8))) +vuint16m8_t vwmulu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m8_m))) +vuint16m8_t vwmulu(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32mf2))) +vuint32mf2_t vwmulu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32mf2_m))) +vuint32mf2_t vwmulu(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m1))) +vuint32m1_t vwmulu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m1_m))) +vuint32m1_t vwmulu(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m2))) +vuint32m2_t vwmulu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m2_m))) +vuint32m2_t vwmulu(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m4))) +vuint32m4_t vwmulu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m4_m))) +vuint32m4_t vwmulu(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m8))) +vuint32m8_t vwmulu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m8_m))) +vuint32m8_t vwmulu(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m1))) +vuint64m1_t vwmulu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m1_m))) +vuint64m1_t vwmulu(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m2))) +vuint64m2_t vwmulu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m2_m))) +vuint64m2_t vwmulu(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m4))) +vuint64m4_t vwmulu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m4_m))) +vuint64m4_t vwmulu(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m8))) +vuint64m8_t vwmulu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m8_m))) +vuint64m8_t vwmulu(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16mf4))) +vuint16mf4_t vwmulu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16mf4_m))) +vuint16mf4_t vwmulu(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16mf2))) +vuint16mf2_t vwmulu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16mf2_m))) +vuint16mf2_t vwmulu(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m1))) +vuint16m1_t vwmulu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m1_m))) +vuint16m1_t vwmulu(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m2))) +vuint16m2_t vwmulu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m2_m))) +vuint16m2_t vwmulu(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m4))) +vuint16m4_t vwmulu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m4_m))) +vuint16m4_t vwmulu(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m8))) +vuint16m8_t vwmulu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m8_m))) +vuint16m8_t vwmulu(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32mf2))) +vuint32mf2_t vwmulu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32mf2_m))) +vuint32mf2_t vwmulu(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m1))) +vuint32m1_t vwmulu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m1_m))) +vuint32m1_t vwmulu(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m2))) +vuint32m2_t vwmulu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m2_m))) +vuint32m2_t vwmulu(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m4))) +vuint32m4_t vwmulu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m4_m))) +vuint32m4_t vwmulu(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m8))) +vuint32m8_t vwmulu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m8_m))) +vuint32m8_t vwmulu(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m1))) +vuint64m1_t vwmulu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m1_m))) +vuint64m1_t vwmulu(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m2))) +vuint64m2_t vwmulu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m2_m))) +vuint64m2_t vwmulu(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m4))) +vuint64m4_t vwmulu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m4_m))) +vuint64m4_t vwmulu(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m8))) +vuint64m8_t vwmulu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m8_m))) +vuint64m8_t vwmulu(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16mf4))) +vint16mf4_t vwmulsu(vint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16mf4_m))) +vint16mf4_t vwmulsu(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16mf2))) +vint16mf2_t vwmulsu(vint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16mf2_m))) +vint16mf2_t vwmulsu(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m1))) +vint16m1_t vwmulsu(vint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m1_m))) +vint16m1_t vwmulsu(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m2))) +vint16m2_t vwmulsu(vint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m2_m))) +vint16m2_t vwmulsu(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m4))) +vint16m4_t vwmulsu(vint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m4_m))) +vint16m4_t vwmulsu(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m8))) +vint16m8_t vwmulsu(vint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m8_m))) +vint16m8_t vwmulsu(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32mf2))) +vint32mf2_t vwmulsu(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32mf2_m))) +vint32mf2_t vwmulsu(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m1))) +vint32m1_t vwmulsu(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m1_m))) +vint32m1_t vwmulsu(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m2))) +vint32m2_t vwmulsu(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m2_m))) +vint32m2_t vwmulsu(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m4))) +vint32m4_t vwmulsu(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m4_m))) +vint32m4_t vwmulsu(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m8))) +vint32m8_t vwmulsu(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m8_m))) +vint32m8_t vwmulsu(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m1))) +vint64m1_t vwmulsu(vint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m1_m))) +vint64m1_t vwmulsu(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m2))) +vint64m2_t vwmulsu(vint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m2_m))) +vint64m2_t vwmulsu(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m4))) +vint64m4_t vwmulsu(vint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m4_m))) +vint64m4_t vwmulsu(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m8))) +vint64m8_t vwmulsu(vint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m8_m))) +vint64m8_t vwmulsu(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16mf4))) +vint16mf4_t vwmulsu(vint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16mf4_m))) +vint16mf4_t vwmulsu(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16mf2))) +vint16mf2_t vwmulsu(vint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16mf2_m))) +vint16mf2_t vwmulsu(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m1))) +vint16m1_t vwmulsu(vint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m1_m))) +vint16m1_t vwmulsu(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m2))) +vint16m2_t vwmulsu(vint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m2_m))) +vint16m2_t vwmulsu(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m4))) +vint16m4_t vwmulsu(vint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m4_m))) +vint16m4_t vwmulsu(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m8))) +vint16m8_t vwmulsu(vint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m8_m))) +vint16m8_t vwmulsu(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32mf2))) +vint32mf2_t vwmulsu(vint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32mf2_m))) +vint32mf2_t vwmulsu(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m1))) +vint32m1_t vwmulsu(vint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m1_m))) +vint32m1_t vwmulsu(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m2))) +vint32m2_t vwmulsu(vint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m2_m))) +vint32m2_t vwmulsu(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m4))) +vint32m4_t vwmulsu(vint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m4_m))) +vint32m4_t vwmulsu(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m8))) +vint32m8_t vwmulsu(vint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m8_m))) +vint32m8_t vwmulsu(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m1))) +vint64m1_t vwmulsu(vint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m1_m))) +vint64m1_t vwmulsu(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m2))) +vint64m2_t vwmulsu(vint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m2_m))) +vint64m2_t vwmulsu(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m4))) +vint64m4_t vwmulsu(vint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m4_m))) +vint64m4_t vwmulsu(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m8))) +vint64m8_t vwmulsu(vint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m8_m))) +vint64m8_t vwmulsu(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m1))) +vint8m1_t vmacc(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m1_m))) +vint8m1_t vmacc(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m2))) +vint8m2_t vmacc(vint8m2_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m2_m))) +vint8m2_t vmacc(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m4))) +vint8m4_t vmacc(vint8m4_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m4_m))) +vint8m4_t vmacc(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m8))) +vint8m8_t vmacc(vint8m8_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m8_m))) +vint8m8_t vmacc(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8mf2))) +vint8mf2_t vmacc(vint8mf2_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8mf2_m))) +vint8mf2_t vmacc(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8mf4))) +vint8mf4_t vmacc(vint8mf4_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8mf4_m))) +vint8mf4_t vmacc(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8mf8))) +vint8mf8_t vmacc(vint8mf8_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8mf8_m))) +vint8mf8_t vmacc(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m1))) +vint16m1_t vmacc(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m1_m))) +vint16m1_t vmacc(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m2))) +vint16m2_t vmacc(vint16m2_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m2_m))) +vint16m2_t vmacc(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m4))) +vint16m4_t vmacc(vint16m4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m4_m))) +vint16m4_t vmacc(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m8))) +vint16m8_t vmacc(vint16m8_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m8_m))) +vint16m8_t vmacc(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16mf2))) +vint16mf2_t vmacc(vint16mf2_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16mf2_m))) +vint16mf2_t vmacc(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16mf4))) +vint16mf4_t vmacc(vint16mf4_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16mf4_m))) +vint16mf4_t vmacc(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m1))) +vint32m1_t vmacc(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m1_m))) +vint32m1_t vmacc(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m2))) +vint32m2_t vmacc(vint32m2_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m2_m))) +vint32m2_t vmacc(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m4))) +vint32m4_t vmacc(vint32m4_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m4_m))) +vint32m4_t vmacc(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m8))) +vint32m8_t vmacc(vint32m8_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m8_m))) +vint32m8_t vmacc(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32mf2))) +vint32mf2_t vmacc(vint32mf2_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32mf2_m))) +vint32mf2_t vmacc(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m1))) +vint64m1_t vmacc(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m1_m))) +vint64m1_t vmacc(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m2))) +vint64m2_t vmacc(vint64m2_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m2_m))) +vint64m2_t vmacc(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m4))) +vint64m4_t vmacc(vint64m4_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m4_m))) +vint64m4_t vmacc(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m8))) +vint64m8_t vmacc(vint64m8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m8_m))) +vint64m8_t vmacc(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m1))) +vint8m1_t vmacc(vint8m1_t op0, int8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m1_m))) +vint8m1_t vmacc(vbool8_t op0, vint8m1_t op1, int8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m2))) +vint8m2_t vmacc(vint8m2_t op0, int8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m2_m))) +vint8m2_t vmacc(vbool4_t op0, vint8m2_t op1, int8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m4))) +vint8m4_t vmacc(vint8m4_t op0, int8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m4_m))) +vint8m4_t vmacc(vbool2_t op0, vint8m4_t op1, int8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m8))) +vint8m8_t vmacc(vint8m8_t op0, int8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m8_m))) +vint8m8_t vmacc(vbool1_t op0, vint8m8_t op1, int8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8mf2))) +vint8mf2_t vmacc(vint8mf2_t op0, int8_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8mf2_m))) +vint8mf2_t vmacc(vbool16_t op0, vint8mf2_t op1, int8_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8mf4))) +vint8mf4_t vmacc(vint8mf4_t op0, int8_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8mf4_m))) +vint8mf4_t vmacc(vbool32_t op0, vint8mf4_t op1, int8_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8mf8))) +vint8mf8_t vmacc(vint8mf8_t op0, int8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8mf8_m))) +vint8mf8_t vmacc(vbool64_t op0, vint8mf8_t op1, int8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m1))) +vint16m1_t vmacc(vint16m1_t op0, int16_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m1_m))) +vint16m1_t vmacc(vbool16_t op0, vint16m1_t op1, int16_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m2))) +vint16m2_t vmacc(vint16m2_t op0, int16_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m2_m))) +vint16m2_t vmacc(vbool8_t op0, vint16m2_t op1, int16_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m4))) +vint16m4_t vmacc(vint16m4_t op0, int16_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m4_m))) +vint16m4_t vmacc(vbool4_t op0, vint16m4_t op1, int16_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m8))) +vint16m8_t vmacc(vint16m8_t op0, int16_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m8_m))) +vint16m8_t vmacc(vbool2_t op0, vint16m8_t op1, int16_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16mf2))) +vint16mf2_t vmacc(vint16mf2_t op0, int16_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16mf2_m))) +vint16mf2_t vmacc(vbool32_t op0, vint16mf2_t op1, int16_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16mf4))) +vint16mf4_t vmacc(vint16mf4_t op0, int16_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16mf4_m))) +vint16mf4_t vmacc(vbool64_t op0, vint16mf4_t op1, int16_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m1))) +vint32m1_t vmacc(vint32m1_t op0, int32_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m1_m))) +vint32m1_t vmacc(vbool32_t op0, vint32m1_t op1, int32_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m2))) +vint32m2_t vmacc(vint32m2_t op0, int32_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m2_m))) +vint32m2_t vmacc(vbool16_t op0, vint32m2_t op1, int32_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m4))) +vint32m4_t vmacc(vint32m4_t op0, int32_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m4_m))) +vint32m4_t vmacc(vbool8_t op0, vint32m4_t op1, int32_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m8))) +vint32m8_t vmacc(vint32m8_t op0, int32_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m8_m))) +vint32m8_t vmacc(vbool4_t op0, vint32m8_t op1, int32_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32mf2))) +vint32mf2_t vmacc(vint32mf2_t op0, int32_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32mf2_m))) +vint32mf2_t vmacc(vbool64_t op0, vint32mf2_t op1, int32_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m1))) +vint64m1_t vmacc(vint64m1_t op0, int64_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m1_m))) +vint64m1_t vmacc(vbool64_t op0, vint64m1_t op1, int64_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m2))) +vint64m2_t vmacc(vint64m2_t op0, int64_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m2_m))) +vint64m2_t vmacc(vbool32_t op0, vint64m2_t op1, int64_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m4))) +vint64m4_t vmacc(vint64m4_t op0, int64_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m4_m))) +vint64m4_t vmacc(vbool16_t op0, vint64m4_t op1, int64_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m8))) +vint64m8_t vmacc(vint64m8_t op0, int64_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m8_m))) +vint64m8_t vmacc(vbool8_t op0, vint64m8_t op1, int64_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m1))) +vuint8m1_t vmacc(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m1_m))) +vuint8m1_t vmacc(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m2))) +vuint8m2_t vmacc(vuint8m2_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m2_m))) +vuint8m2_t vmacc(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m4))) +vuint8m4_t vmacc(vuint8m4_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m4_m))) +vuint8m4_t vmacc(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m8))) +vuint8m8_t vmacc(vuint8m8_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m8_m))) +vuint8m8_t vmacc(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8mf2))) +vuint8mf2_t vmacc(vuint8mf2_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8mf2_m))) +vuint8mf2_t vmacc(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8mf4))) +vuint8mf4_t vmacc(vuint8mf4_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8mf4_m))) +vuint8mf4_t vmacc(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8mf8))) +vuint8mf8_t vmacc(vuint8mf8_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8mf8_m))) +vuint8mf8_t vmacc(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m1))) +vuint16m1_t vmacc(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m1_m))) +vuint16m1_t vmacc(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m2))) +vuint16m2_t vmacc(vuint16m2_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m2_m))) +vuint16m2_t vmacc(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m4))) +vuint16m4_t vmacc(vuint16m4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m4_m))) +vuint16m4_t vmacc(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m8))) +vuint16m8_t vmacc(vuint16m8_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m8_m))) +vuint16m8_t vmacc(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16mf2))) +vuint16mf2_t vmacc(vuint16mf2_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16mf2_m))) +vuint16mf2_t vmacc(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16mf4))) +vuint16mf4_t vmacc(vuint16mf4_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16mf4_m))) +vuint16mf4_t vmacc(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m1))) +vuint32m1_t vmacc(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m1_m))) +vuint32m1_t vmacc(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m2))) +vuint32m2_t vmacc(vuint32m2_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m2_m))) +vuint32m2_t vmacc(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m4))) +vuint32m4_t vmacc(vuint32m4_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m4_m))) +vuint32m4_t vmacc(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m8))) +vuint32m8_t vmacc(vuint32m8_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m8_m))) +vuint32m8_t vmacc(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32mf2))) +vuint32mf2_t vmacc(vuint32mf2_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32mf2_m))) +vuint32mf2_t vmacc(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m1))) +vuint64m1_t vmacc(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m1_m))) +vuint64m1_t vmacc(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m2))) +vuint64m2_t vmacc(vuint64m2_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m2_m))) +vuint64m2_t vmacc(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m4))) +vuint64m4_t vmacc(vuint64m4_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m4_m))) +vuint64m4_t vmacc(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m8))) +vuint64m8_t vmacc(vuint64m8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m8_m))) +vuint64m8_t vmacc(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m1))) +vuint8m1_t vmacc(vuint8m1_t op0, uint8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m1_m))) +vuint8m1_t vmacc(vbool8_t op0, vuint8m1_t op1, uint8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m2))) +vuint8m2_t vmacc(vuint8m2_t op0, uint8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m2_m))) +vuint8m2_t vmacc(vbool4_t op0, vuint8m2_t op1, uint8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m4))) +vuint8m4_t vmacc(vuint8m4_t op0, uint8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m4_m))) +vuint8m4_t vmacc(vbool2_t op0, vuint8m4_t op1, uint8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m8))) +vuint8m8_t vmacc(vuint8m8_t op0, uint8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m8_m))) +vuint8m8_t vmacc(vbool1_t op0, vuint8m8_t op1, uint8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8mf2))) +vuint8mf2_t vmacc(vuint8mf2_t op0, uint8_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8mf2_m))) +vuint8mf2_t vmacc(vbool16_t op0, vuint8mf2_t op1, uint8_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8mf4))) +vuint8mf4_t vmacc(vuint8mf4_t op0, uint8_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8mf4_m))) +vuint8mf4_t vmacc(vbool32_t op0, vuint8mf4_t op1, uint8_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8mf8))) +vuint8mf8_t vmacc(vuint8mf8_t op0, uint8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8mf8_m))) +vuint8mf8_t vmacc(vbool64_t op0, vuint8mf8_t op1, uint8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m1))) +vuint16m1_t vmacc(vuint16m1_t op0, uint16_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m1_m))) +vuint16m1_t vmacc(vbool16_t op0, vuint16m1_t op1, uint16_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m2))) +vuint16m2_t vmacc(vuint16m2_t op0, uint16_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m2_m))) +vuint16m2_t vmacc(vbool8_t op0, vuint16m2_t op1, uint16_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m4))) +vuint16m4_t vmacc(vuint16m4_t op0, uint16_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m4_m))) +vuint16m4_t vmacc(vbool4_t op0, vuint16m4_t op1, uint16_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m8))) +vuint16m8_t vmacc(vuint16m8_t op0, uint16_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m8_m))) +vuint16m8_t vmacc(vbool2_t op0, vuint16m8_t op1, uint16_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16mf2))) +vuint16mf2_t vmacc(vuint16mf2_t op0, uint16_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16mf2_m))) +vuint16mf2_t vmacc(vbool32_t op0, vuint16mf2_t op1, uint16_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16mf4))) +vuint16mf4_t vmacc(vuint16mf4_t op0, uint16_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16mf4_m))) +vuint16mf4_t vmacc(vbool64_t op0, vuint16mf4_t op1, uint16_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m1))) +vuint32m1_t vmacc(vuint32m1_t op0, uint32_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m1_m))) +vuint32m1_t vmacc(vbool32_t op0, vuint32m1_t op1, uint32_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m2))) +vuint32m2_t vmacc(vuint32m2_t op0, uint32_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m2_m))) +vuint32m2_t vmacc(vbool16_t op0, vuint32m2_t op1, uint32_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m4))) +vuint32m4_t vmacc(vuint32m4_t op0, uint32_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m4_m))) +vuint32m4_t vmacc(vbool8_t op0, vuint32m4_t op1, uint32_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m8))) +vuint32m8_t vmacc(vuint32m8_t op0, uint32_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m8_m))) +vuint32m8_t vmacc(vbool4_t op0, vuint32m8_t op1, uint32_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32mf2))) +vuint32mf2_t vmacc(vuint32mf2_t op0, uint32_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32mf2_m))) +vuint32mf2_t vmacc(vbool64_t op0, vuint32mf2_t op1, uint32_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m1))) +vuint64m1_t vmacc(vuint64m1_t op0, uint64_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m1_m))) +vuint64m1_t vmacc(vbool64_t op0, vuint64m1_t op1, uint64_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m2))) +vuint64m2_t vmacc(vuint64m2_t op0, uint64_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m2_m))) +vuint64m2_t vmacc(vbool32_t op0, vuint64m2_t op1, uint64_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m4))) +vuint64m4_t vmacc(vuint64m4_t op0, uint64_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m4_m))) +vuint64m4_t vmacc(vbool16_t op0, vuint64m4_t op1, uint64_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m8))) +vuint64m8_t vmacc(vuint64m8_t op0, uint64_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m8_m))) +vuint64m8_t vmacc(vbool8_t op0, vuint64m8_t op1, uint64_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m1))) +vint8m1_t vnmsac(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m1_m))) +vint8m1_t vnmsac(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m2))) +vint8m2_t vnmsac(vint8m2_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m2_m))) +vint8m2_t vnmsac(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m4))) +vint8m4_t vnmsac(vint8m4_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m4_m))) +vint8m4_t vnmsac(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m8))) +vint8m8_t vnmsac(vint8m8_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m8_m))) +vint8m8_t vnmsac(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8mf2))) +vint8mf2_t vnmsac(vint8mf2_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8mf2_m))) +vint8mf2_t vnmsac(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8mf4))) +vint8mf4_t vnmsac(vint8mf4_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8mf4_m))) +vint8mf4_t vnmsac(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8mf8))) +vint8mf8_t vnmsac(vint8mf8_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8mf8_m))) +vint8mf8_t vnmsac(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m1))) +vint16m1_t vnmsac(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m1_m))) +vint16m1_t vnmsac(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m2))) +vint16m2_t vnmsac(vint16m2_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m2_m))) +vint16m2_t vnmsac(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m4))) +vint16m4_t vnmsac(vint16m4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m4_m))) +vint16m4_t vnmsac(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m8))) +vint16m8_t vnmsac(vint16m8_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m8_m))) +vint16m8_t vnmsac(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16mf2))) +vint16mf2_t vnmsac(vint16mf2_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16mf2_m))) +vint16mf2_t vnmsac(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16mf4))) +vint16mf4_t vnmsac(vint16mf4_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16mf4_m))) +vint16mf4_t vnmsac(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m1))) +vint32m1_t vnmsac(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m1_m))) +vint32m1_t vnmsac(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m2))) +vint32m2_t vnmsac(vint32m2_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m2_m))) +vint32m2_t vnmsac(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m4))) +vint32m4_t vnmsac(vint32m4_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m4_m))) +vint32m4_t vnmsac(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m8))) +vint32m8_t vnmsac(vint32m8_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m8_m))) +vint32m8_t vnmsac(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32mf2))) +vint32mf2_t vnmsac(vint32mf2_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32mf2_m))) +vint32mf2_t vnmsac(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m1))) +vint64m1_t vnmsac(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m1_m))) +vint64m1_t vnmsac(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m2))) +vint64m2_t vnmsac(vint64m2_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m2_m))) +vint64m2_t vnmsac(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m4))) +vint64m4_t vnmsac(vint64m4_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m4_m))) +vint64m4_t vnmsac(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m8))) +vint64m8_t vnmsac(vint64m8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m8_m))) +vint64m8_t vnmsac(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m1))) +void vsuxei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m1_m))) +void vsuxei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m2))) +void vsuxei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m2_m))) +void vsuxei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m4))) +void vsuxei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m4_m))) +void vsuxei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m8))) +void vsuxei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m8_m))) +void vsuxei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32mf2))) +void vsuxei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32mf2_m))) +void vsuxei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m1))) +vint8m1_t vnmsac(vint8m1_t op0, int8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m1_m))) +vint8m1_t vnmsac(vbool8_t op0, vint8m1_t op1, int8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m2))) +vint8m2_t vnmsac(vint8m2_t op0, int8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m2_m))) +vint8m2_t vnmsac(vbool4_t op0, vint8m2_t op1, int8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m4))) +vint8m4_t vnmsac(vint8m4_t op0, int8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m4_m))) +vint8m4_t vnmsac(vbool2_t op0, vint8m4_t op1, int8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m8))) +vint8m8_t vnmsac(vint8m8_t op0, int8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m8_m))) +vint8m8_t vnmsac(vbool1_t op0, vint8m8_t op1, int8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8mf2))) +vint8mf2_t vnmsac(vint8mf2_t op0, int8_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8mf2_m))) +vint8mf2_t vnmsac(vbool16_t op0, vint8mf2_t op1, int8_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8mf4))) +vint8mf4_t vnmsac(vint8mf4_t op0, int8_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8mf4_m))) +vint8mf4_t vnmsac(vbool32_t op0, vint8mf4_t op1, int8_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8mf8))) +vint8mf8_t vnmsac(vint8mf8_t op0, int8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8mf8_m))) +vint8mf8_t vnmsac(vbool64_t op0, vint8mf8_t op1, int8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m1))) +vint16m1_t vnmsac(vint16m1_t op0, int16_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m1_m))) +vint16m1_t vnmsac(vbool16_t op0, vint16m1_t op1, int16_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m2))) +vint16m2_t vnmsac(vint16m2_t op0, int16_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m2_m))) +vint16m2_t vnmsac(vbool8_t op0, vint16m2_t op1, int16_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m4))) +vint16m4_t vnmsac(vint16m4_t op0, int16_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m4_m))) +vint16m4_t vnmsac(vbool4_t op0, vint16m4_t op1, int16_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m8))) +vint16m8_t vnmsac(vint16m8_t op0, int16_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m8_m))) +vint16m8_t vnmsac(vbool2_t op0, vint16m8_t op1, int16_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16mf2))) +vint16mf2_t vnmsac(vint16mf2_t op0, int16_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16mf2_m))) +vint16mf2_t vnmsac(vbool32_t op0, vint16mf2_t op1, int16_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16mf4))) +vint16mf4_t vnmsac(vint16mf4_t op0, int16_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16mf4_m))) +vint16mf4_t vnmsac(vbool64_t op0, vint16mf4_t op1, int16_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m1))) +vint32m1_t vnmsac(vint32m1_t op0, int32_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m1_m))) +vint32m1_t vnmsac(vbool32_t op0, vint32m1_t op1, int32_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m2))) +vint32m2_t vnmsac(vint32m2_t op0, int32_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m2_m))) +vint32m2_t vnmsac(vbool16_t op0, vint32m2_t op1, int32_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m4))) +vint32m4_t vnmsac(vint32m4_t op0, int32_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m4_m))) +vint32m4_t vnmsac(vbool8_t op0, vint32m4_t op1, int32_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m8))) +vint32m8_t vnmsac(vint32m8_t op0, int32_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m8_m))) +vint32m8_t vnmsac(vbool4_t op0, vint32m8_t op1, int32_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32mf2))) +vint32mf2_t vnmsac(vint32mf2_t op0, int32_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32mf2_m))) +vint32mf2_t vnmsac(vbool64_t op0, vint32mf2_t op1, int32_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m1))) +vint64m1_t vnmsac(vint64m1_t op0, int64_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m1_m))) +vint64m1_t vnmsac(vbool64_t op0, vint64m1_t op1, int64_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m2))) +vint64m2_t vnmsac(vint64m2_t op0, int64_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m2_m))) +vint64m2_t vnmsac(vbool32_t op0, vint64m2_t op1, int64_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m4))) +vint64m4_t vnmsac(vint64m4_t op0, int64_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m4_m))) +vint64m4_t vnmsac(vbool16_t op0, vint64m4_t op1, int64_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m8))) +vint64m8_t vnmsac(vint64m8_t op0, int64_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m8_m))) +vint64m8_t vnmsac(vbool8_t op0, vint64m8_t op1, int64_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m1))) +vuint8m1_t vnmsac(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m1_m))) +vuint8m1_t vnmsac(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m2))) +vuint8m2_t vnmsac(vuint8m2_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m2_m))) +vuint8m2_t vnmsac(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m4))) +vuint8m4_t vnmsac(vuint8m4_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m4_m))) +vuint8m4_t vnmsac(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m8))) +vuint8m8_t vnmsac(vuint8m8_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m8_m))) +vuint8m8_t vnmsac(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8mf2))) +vuint8mf2_t vnmsac(vuint8mf2_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8mf2_m))) +vuint8mf2_t vnmsac(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8mf4))) +vuint8mf4_t vnmsac(vuint8mf4_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8mf4_m))) +vuint8mf4_t vnmsac(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8mf8))) +vuint8mf8_t vnmsac(vuint8mf8_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8mf8_m))) +vuint8mf8_t vnmsac(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m1))) +vuint16m1_t vnmsac(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m1_m))) +vuint16m1_t vnmsac(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m2))) +vuint16m2_t vnmsac(vuint16m2_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m2_m))) +vuint16m2_t vnmsac(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m4))) +vuint16m4_t vnmsac(vuint16m4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m4_m))) +vuint16m4_t vnmsac(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m8))) +vuint16m8_t vnmsac(vuint16m8_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m8_m))) +vuint16m8_t vnmsac(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16mf2))) +vuint16mf2_t vnmsac(vuint16mf2_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16mf2_m))) +vuint16mf2_t vnmsac(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16mf4))) +vuint16mf4_t vnmsac(vuint16mf4_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16mf4_m))) +vuint16mf4_t vnmsac(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m1))) +vuint32m1_t vnmsac(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m1_m))) +vuint32m1_t vnmsac(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m2))) +vuint32m2_t vnmsac(vuint32m2_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m2_m))) +vuint32m2_t vnmsac(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m4))) +vuint32m4_t vnmsac(vuint32m4_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m4_m))) +vuint32m4_t vnmsac(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m8))) +vuint32m8_t vnmsac(vuint32m8_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m8_m))) +vuint32m8_t vnmsac(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32mf2))) +vuint32mf2_t vnmsac(vuint32mf2_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32mf2_m))) +vuint32mf2_t vnmsac(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m1))) +vuint64m1_t vnmsac(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m1_m))) +vuint64m1_t vnmsac(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m2))) +vuint64m2_t vnmsac(vuint64m2_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m2_m))) +vuint64m2_t vnmsac(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m4))) +vuint64m4_t vnmsac(vuint64m4_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m4_m))) +vuint64m4_t vnmsac(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m8))) +vuint64m8_t vnmsac(vuint64m8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m8_m))) +vuint64m8_t vnmsac(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m1))) +vuint8m1_t vnmsac(vuint8m1_t op0, uint8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m1_m))) +vuint8m1_t vnmsac(vbool8_t op0, vuint8m1_t op1, uint8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m2))) +vuint8m2_t vnmsac(vuint8m2_t op0, uint8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m2_m))) +vuint8m2_t vnmsac(vbool4_t op0, vuint8m2_t op1, uint8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m4))) +vuint8m4_t vnmsac(vuint8m4_t op0, uint8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m4_m))) +vuint8m4_t vnmsac(vbool2_t op0, vuint8m4_t op1, uint8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m8))) +vuint8m8_t vnmsac(vuint8m8_t op0, uint8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m8_m))) +vuint8m8_t vnmsac(vbool1_t op0, vuint8m8_t op1, uint8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8mf2))) +vuint8mf2_t vnmsac(vuint8mf2_t op0, uint8_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8mf2_m))) +vuint8mf2_t vnmsac(vbool16_t op0, vuint8mf2_t op1, uint8_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8mf4))) +vuint8mf4_t vnmsac(vuint8mf4_t op0, uint8_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8mf4_m))) +vuint8mf4_t vnmsac(vbool32_t op0, vuint8mf4_t op1, uint8_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8mf8))) +vuint8mf8_t vnmsac(vuint8mf8_t op0, uint8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8mf8_m))) +vuint8mf8_t vnmsac(vbool64_t op0, vuint8mf8_t op1, uint8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m1))) +vuint16m1_t vnmsac(vuint16m1_t op0, uint16_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m1_m))) +vuint16m1_t vnmsac(vbool16_t op0, vuint16m1_t op1, uint16_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m2))) +vuint16m2_t vnmsac(vuint16m2_t op0, uint16_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m2_m))) +vuint16m2_t vnmsac(vbool8_t op0, vuint16m2_t op1, uint16_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m4))) +vuint16m4_t vnmsac(vuint16m4_t op0, uint16_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m4_m))) +vuint16m4_t vnmsac(vbool4_t op0, vuint16m4_t op1, uint16_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m8))) +vuint16m8_t vnmsac(vuint16m8_t op0, uint16_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m8_m))) +vuint16m8_t vnmsac(vbool2_t op0, vuint16m8_t op1, uint16_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16mf2))) +vuint16mf2_t vnmsac(vuint16mf2_t op0, uint16_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16mf2_m))) +vuint16mf2_t vnmsac(vbool32_t op0, vuint16mf2_t op1, uint16_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16mf4))) +vuint16mf4_t vnmsac(vuint16mf4_t op0, uint16_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16mf4_m))) +vuint16mf4_t vnmsac(vbool64_t op0, vuint16mf4_t op1, uint16_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m1))) +vuint32m1_t vnmsac(vuint32m1_t op0, uint32_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m1_m))) +vuint32m1_t vnmsac(vbool32_t op0, vuint32m1_t op1, uint32_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m2))) +vuint32m2_t vnmsac(vuint32m2_t op0, uint32_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m2_m))) +vuint32m2_t vnmsac(vbool16_t op0, vuint32m2_t op1, uint32_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m4))) +vuint32m4_t vnmsac(vuint32m4_t op0, uint32_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m4_m))) +vuint32m4_t vnmsac(vbool8_t op0, vuint32m4_t op1, uint32_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m8))) +vuint32m8_t vnmsac(vuint32m8_t op0, uint32_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m8_m))) +vuint32m8_t vnmsac(vbool4_t op0, vuint32m8_t op1, uint32_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32mf2))) +vuint32mf2_t vnmsac(vuint32mf2_t op0, uint32_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32mf2_m))) +vuint32mf2_t vnmsac(vbool64_t op0, vuint32mf2_t op1, uint32_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m1))) +vuint64m1_t vnmsac(vuint64m1_t op0, uint64_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m1_m))) +vuint64m1_t vnmsac(vbool64_t op0, vuint64m1_t op1, uint64_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m2))) +vuint64m2_t vnmsac(vuint64m2_t op0, uint64_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m2_m))) +vuint64m2_t vnmsac(vbool32_t op0, vuint64m2_t op1, uint64_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m4))) +vuint64m4_t vnmsac(vuint64m4_t op0, uint64_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m4_m))) +vuint64m4_t vnmsac(vbool16_t op0, vuint64m4_t op1, uint64_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m8))) +vuint64m8_t vnmsac(vuint64m8_t op0, uint64_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m8_m))) +vuint64m8_t vnmsac(vbool8_t op0, vuint64m8_t op1, uint64_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m1))) +vint8m1_t vmadd(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m1_m))) +vint8m1_t vmadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m2))) +vint8m2_t vmadd(vint8m2_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m2_m))) +vint8m2_t vmadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m4))) +vint8m4_t vmadd(vint8m4_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m4_m))) +vint8m4_t vmadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m8))) +vint8m8_t vmadd(vint8m8_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m8_m))) +vint8m8_t vmadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8mf2))) +vint8mf2_t vmadd(vint8mf2_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8mf2_m))) +vint8mf2_t vmadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8mf4))) +vint8mf4_t vmadd(vint8mf4_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8mf4_m))) +vint8mf4_t vmadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8mf8))) +vint8mf8_t vmadd(vint8mf8_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8mf8_m))) +vint8mf8_t vmadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m1))) +vint16m1_t vmadd(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m1_m))) +vint16m1_t vmadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m2))) +vint16m2_t vmadd(vint16m2_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m2_m))) +vint16m2_t vmadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m4))) +vint16m4_t vmadd(vint16m4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m4_m))) +vint16m4_t vmadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m8))) +vint16m8_t vmadd(vint16m8_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m8_m))) +vint16m8_t vmadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16mf2))) +vint16mf2_t vmadd(vint16mf2_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16mf2_m))) +vint16mf2_t vmadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16mf4))) +vint16mf4_t vmadd(vint16mf4_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16mf4_m))) +vint16mf4_t vmadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m1))) +vint32m1_t vmadd(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m1_m))) +vint32m1_t vmadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m2))) +vint32m2_t vmadd(vint32m2_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m2_m))) +vint32m2_t vmadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m4))) +vint32m4_t vmadd(vint32m4_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m4_m))) +vint32m4_t vmadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m8))) +vint32m8_t vmadd(vint32m8_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m8_m))) +vint32m8_t vmadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32mf2))) +vint32mf2_t vmadd(vint32mf2_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32mf2_m))) +vint32mf2_t vmadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m1))) +vint64m1_t vmadd(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m1_m))) +vint64m1_t vmadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m2))) +vint64m2_t vmadd(vint64m2_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m2_m))) +vint64m2_t vmadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m4))) +vint64m4_t vmadd(vint64m4_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m4_m))) +vint64m4_t vmadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m8))) +vint64m8_t vmadd(vint64m8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m8_m))) +vint64m8_t vmadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m1))) +vint8m1_t vmadd(vint8m1_t op0, int8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m1_m))) +vint8m1_t vmadd(vbool8_t op0, vint8m1_t op1, int8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m2))) +vint8m2_t vmadd(vint8m2_t op0, int8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m2_m))) +vint8m2_t vmadd(vbool4_t op0, vint8m2_t op1, int8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m4))) +vint8m4_t vmadd(vint8m4_t op0, int8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m4_m))) +vint8m4_t vmadd(vbool2_t op0, vint8m4_t op1, int8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m8))) +vint8m8_t vmadd(vint8m8_t op0, int8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m8_m))) +vint8m8_t vmadd(vbool1_t op0, vint8m8_t op1, int8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8mf2))) +vint8mf2_t vmadd(vint8mf2_t op0, int8_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8mf2_m))) +vint8mf2_t vmadd(vbool16_t op0, vint8mf2_t op1, int8_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8mf4))) +vint8mf4_t vmadd(vint8mf4_t op0, int8_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8mf4_m))) +vint8mf4_t vmadd(vbool32_t op0, vint8mf4_t op1, int8_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8mf8))) +vint8mf8_t vmadd(vint8mf8_t op0, int8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8mf8_m))) +vint8mf8_t vmadd(vbool64_t op0, vint8mf8_t op1, int8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m1))) +vint16m1_t vmadd(vint16m1_t op0, int16_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m1_m))) +vint16m1_t vmadd(vbool16_t op0, vint16m1_t op1, int16_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m2))) +vint16m2_t vmadd(vint16m2_t op0, int16_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m2_m))) +vint16m2_t vmadd(vbool8_t op0, vint16m2_t op1, int16_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m4))) +vint16m4_t vmadd(vint16m4_t op0, int16_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m4_m))) +vint16m4_t vmadd(vbool4_t op0, vint16m4_t op1, int16_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m8))) +vint16m8_t vmadd(vint16m8_t op0, int16_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m8_m))) +vint16m8_t vmadd(vbool2_t op0, vint16m8_t op1, int16_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16mf2))) +vint16mf2_t vmadd(vint16mf2_t op0, int16_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16mf2_m))) +vint16mf2_t vmadd(vbool32_t op0, vint16mf2_t op1, int16_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16mf4))) +vint16mf4_t vmadd(vint16mf4_t op0, int16_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16mf4_m))) +vint16mf4_t vmadd(vbool64_t op0, vint16mf4_t op1, int16_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m1))) +vint32m1_t vmadd(vint32m1_t op0, int32_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m1_m))) +vint32m1_t vmadd(vbool32_t op0, vint32m1_t op1, int32_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m2))) +vint32m2_t vmadd(vint32m2_t op0, int32_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m2_m))) +vint32m2_t vmadd(vbool16_t op0, vint32m2_t op1, int32_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m4))) +vint32m4_t vmadd(vint32m4_t op0, int32_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m4_m))) +vint32m4_t vmadd(vbool8_t op0, vint32m4_t op1, int32_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m8))) +vint32m8_t vmadd(vint32m8_t op0, int32_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m8_m))) +vint32m8_t vmadd(vbool4_t op0, vint32m8_t op1, int32_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32mf2))) +vint32mf2_t vmadd(vint32mf2_t op0, int32_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32mf2_m))) +vint32mf2_t vmadd(vbool64_t op0, vint32mf2_t op1, int32_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m1))) +vint64m1_t vmadd(vint64m1_t op0, int64_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m1_m))) +vint64m1_t vmadd(vbool64_t op0, vint64m1_t op1, int64_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m2))) +vint64m2_t vmadd(vint64m2_t op0, int64_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m2_m))) +vint64m2_t vmadd(vbool32_t op0, vint64m2_t op1, int64_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m4))) +vint64m4_t vmadd(vint64m4_t op0, int64_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m4_m))) +vint64m4_t vmadd(vbool16_t op0, vint64m4_t op1, int64_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m8))) +vint64m8_t vmadd(vint64m8_t op0, int64_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m8_m))) +vint64m8_t vmadd(vbool8_t op0, vint64m8_t op1, int64_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m1))) +vuint8m1_t vmadd(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m1_m))) +vuint8m1_t vmadd(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m2))) +vuint8m2_t vmadd(vuint8m2_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m2_m))) +vuint8m2_t vmadd(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m4))) +vuint8m4_t vmadd(vuint8m4_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m4_m))) +vuint8m4_t vmadd(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m8))) +vuint8m8_t vmadd(vuint8m8_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m8_m))) +vuint8m8_t vmadd(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8mf2))) +vuint8mf2_t vmadd(vuint8mf2_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8mf2_m))) +vuint8mf2_t vmadd(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8mf4))) +vuint8mf4_t vmadd(vuint8mf4_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8mf4_m))) +vuint8mf4_t vmadd(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8mf8))) +vuint8mf8_t vmadd(vuint8mf8_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8mf8_m))) +vuint8mf8_t vmadd(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m1))) +vuint16m1_t vmadd(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m1_m))) +vuint16m1_t vmadd(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m2))) +vuint16m2_t vmadd(vuint16m2_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m2_m))) +vuint16m2_t vmadd(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m4))) +vuint16m4_t vmadd(vuint16m4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m4_m))) +vuint16m4_t vmadd(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m8))) +vuint16m8_t vmadd(vuint16m8_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m8_m))) +vuint16m8_t vmadd(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16mf2))) +vuint16mf2_t vmadd(vuint16mf2_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16mf2_m))) +vuint16mf2_t vmadd(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16mf4))) +vuint16mf4_t vmadd(vuint16mf4_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16mf4_m))) +vuint16mf4_t vmadd(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m1))) +vuint32m1_t vmadd(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m1_m))) +vuint32m1_t vmadd(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m2))) +vuint32m2_t vmadd(vuint32m2_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m2_m))) +vuint32m2_t vmadd(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m4))) +vuint32m4_t vmadd(vuint32m4_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m4_m))) +vuint32m4_t vmadd(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m8))) +vuint32m8_t vmadd(vuint32m8_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m8_m))) +vuint32m8_t vmadd(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32mf2))) +vuint32mf2_t vmadd(vuint32mf2_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32mf2_m))) +vuint32mf2_t vmadd(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m1))) +vuint64m1_t vmadd(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m1_m))) +vuint64m1_t vmadd(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m2))) +vuint64m2_t vmadd(vuint64m2_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m2_m))) +vuint64m2_t vmadd(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m4))) +vuint64m4_t vmadd(vuint64m4_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m4_m))) +vuint64m4_t vmadd(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m8))) +vuint64m8_t vmadd(vuint64m8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m8_m))) +vuint64m8_t vmadd(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m1))) +vuint8m1_t vmadd(vuint8m1_t op0, uint8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m1_m))) +vuint8m1_t vmadd(vbool8_t op0, vuint8m1_t op1, uint8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m2))) +vuint8m2_t vmadd(vuint8m2_t op0, uint8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m2_m))) +vuint8m2_t vmadd(vbool4_t op0, vuint8m2_t op1, uint8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m4))) +vuint8m4_t vmadd(vuint8m4_t op0, uint8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m4_m))) +vuint8m4_t vmadd(vbool2_t op0, vuint8m4_t op1, uint8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m8))) +vuint8m8_t vmadd(vuint8m8_t op0, uint8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m8_m))) +vuint8m8_t vmadd(vbool1_t op0, vuint8m8_t op1, uint8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8mf2))) +vuint8mf2_t vmadd(vuint8mf2_t op0, uint8_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8mf2_m))) +vuint8mf2_t vmadd(vbool16_t op0, vuint8mf2_t op1, uint8_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8mf4))) +vuint8mf4_t vmadd(vuint8mf4_t op0, uint8_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8mf4_m))) +vuint8mf4_t vmadd(vbool32_t op0, vuint8mf4_t op1, uint8_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8mf8))) +vuint8mf8_t vmadd(vuint8mf8_t op0, uint8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8mf8_m))) +vuint8mf8_t vmadd(vbool64_t op0, vuint8mf8_t op1, uint8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m1))) +vuint16m1_t vmadd(vuint16m1_t op0, uint16_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m1_m))) +vuint16m1_t vmadd(vbool16_t op0, vuint16m1_t op1, uint16_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m2))) +vuint16m2_t vmadd(vuint16m2_t op0, uint16_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m2_m))) +vuint16m2_t vmadd(vbool8_t op0, vuint16m2_t op1, uint16_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m4))) +vuint16m4_t vmadd(vuint16m4_t op0, uint16_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m4_m))) +vuint16m4_t vmadd(vbool4_t op0, vuint16m4_t op1, uint16_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m8))) +vuint16m8_t vmadd(vuint16m8_t op0, uint16_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m8_m))) +vuint16m8_t vmadd(vbool2_t op0, vuint16m8_t op1, uint16_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16mf2))) +vuint16mf2_t vmadd(vuint16mf2_t op0, uint16_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16mf2_m))) +vuint16mf2_t vmadd(vbool32_t op0, vuint16mf2_t op1, uint16_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16mf4))) +vuint16mf4_t vmadd(vuint16mf4_t op0, uint16_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16mf4_m))) +vuint16mf4_t vmadd(vbool64_t op0, vuint16mf4_t op1, uint16_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m1))) +vuint32m1_t vmadd(vuint32m1_t op0, uint32_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m1_m))) +vuint32m1_t vmadd(vbool32_t op0, vuint32m1_t op1, uint32_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m2))) +vuint32m2_t vmadd(vuint32m2_t op0, uint32_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m2_m))) +vuint32m2_t vmadd(vbool16_t op0, vuint32m2_t op1, uint32_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m4))) +vuint32m4_t vmadd(vuint32m4_t op0, uint32_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m4_m))) +vuint32m4_t vmadd(vbool8_t op0, vuint32m4_t op1, uint32_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m8))) +vuint32m8_t vmadd(vuint32m8_t op0, uint32_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m8_m))) +vuint32m8_t vmadd(vbool4_t op0, vuint32m8_t op1, uint32_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32mf2))) +vuint32mf2_t vmadd(vuint32mf2_t op0, uint32_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32mf2_m))) +vuint32mf2_t vmadd(vbool64_t op0, vuint32mf2_t op1, uint32_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m1))) +vuint64m1_t vmadd(vuint64m1_t op0, uint64_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m1_m))) +vuint64m1_t vmadd(vbool64_t op0, vuint64m1_t op1, uint64_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m2))) +vuint64m2_t vmadd(vuint64m2_t op0, uint64_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m2_m))) +vuint64m2_t vmadd(vbool32_t op0, vuint64m2_t op1, uint64_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m4))) +vuint64m4_t vmadd(vuint64m4_t op0, uint64_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m4_m))) +vuint64m4_t vmadd(vbool16_t op0, vuint64m4_t op1, uint64_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m8))) +vuint64m8_t vmadd(vuint64m8_t op0, uint64_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m8_m))) +vuint64m8_t vmadd(vbool8_t op0, vuint64m8_t op1, uint64_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m1))) +vint8m1_t vnmsub(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m1_m))) +vint8m1_t vnmsub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m2))) +vint8m2_t vnmsub(vint8m2_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m2_m))) +vint8m2_t vnmsub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m4))) +vint8m4_t vnmsub(vint8m4_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m4_m))) +vint8m4_t vnmsub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m8))) +vint8m8_t vnmsub(vint8m8_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m8_m))) +vint8m8_t vnmsub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8mf2))) +vint8mf2_t vnmsub(vint8mf2_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8mf2_m))) +vint8mf2_t vnmsub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8mf4))) +vint8mf4_t vnmsub(vint8mf4_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8mf4_m))) +vint8mf4_t vnmsub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8mf8))) +vint8mf8_t vnmsub(vint8mf8_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8mf8_m))) +vint8mf8_t vnmsub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m1))) +vint16m1_t vnmsub(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m1_m))) +vint16m1_t vnmsub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m2))) +vint16m2_t vnmsub(vint16m2_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m2_m))) +vint16m2_t vnmsub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m4))) +vint16m4_t vnmsub(vint16m4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m4_m))) +vint16m4_t vnmsub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m8))) +vint16m8_t vnmsub(vint16m8_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m8_m))) +vint16m8_t vnmsub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16mf2))) +vint16mf2_t vnmsub(vint16mf2_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16mf2_m))) +vint16mf2_t vnmsub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16mf4))) +vint16mf4_t vnmsub(vint16mf4_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16mf4_m))) +vint16mf4_t vnmsub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m1))) +vint32m1_t vnmsub(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m1_m))) +vint32m1_t vnmsub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m2))) +vint32m2_t vnmsub(vint32m2_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m2_m))) +vint32m2_t vnmsub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m4))) +vint32m4_t vnmsub(vint32m4_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m4_m))) +vint32m4_t vnmsub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m8))) +vint32m8_t vnmsub(vint32m8_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m8_m))) +vint32m8_t vnmsub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32mf2))) +vint32mf2_t vnmsub(vint32mf2_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32mf2_m))) +vint32mf2_t vnmsub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m1))) +vint64m1_t vnmsub(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m1_m))) +vint64m1_t vnmsub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m2))) +vint64m2_t vnmsub(vint64m2_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m2_m))) +vint64m2_t vnmsub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m4))) +vint64m4_t vnmsub(vint64m4_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m4_m))) +vint64m4_t vnmsub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m8))) +vint64m8_t vnmsub(vint64m8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m8_m))) +vint64m8_t vnmsub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m1))) +vint8m1_t vnmsub(vint8m1_t op0, int8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m1_m))) +vint8m1_t vnmsub(vbool8_t op0, vint8m1_t op1, int8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m2))) +vint8m2_t vnmsub(vint8m2_t op0, int8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m2_m))) +vint8m2_t vnmsub(vbool4_t op0, vint8m2_t op1, int8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m4))) +vint8m4_t vnmsub(vint8m4_t op0, int8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m4_m))) +vint8m4_t vnmsub(vbool2_t op0, vint8m4_t op1, int8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m8))) +vint8m8_t vnmsub(vint8m8_t op0, int8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m8_m))) +vint8m8_t vnmsub(vbool1_t op0, vint8m8_t op1, int8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8mf2))) +vint8mf2_t vnmsub(vint8mf2_t op0, int8_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8mf2_m))) +vint8mf2_t vnmsub(vbool16_t op0, vint8mf2_t op1, int8_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8mf4))) +vint8mf4_t vnmsub(vint8mf4_t op0, int8_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8mf4_m))) +vint8mf4_t vnmsub(vbool32_t op0, vint8mf4_t op1, int8_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8mf8))) +vint8mf8_t vnmsub(vint8mf8_t op0, int8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8mf8_m))) +vint8mf8_t vnmsub(vbool64_t op0, vint8mf8_t op1, int8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m1))) +vint16m1_t vnmsub(vint16m1_t op0, int16_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m1_m))) +vint16m1_t vnmsub(vbool16_t op0, vint16m1_t op1, int16_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m2))) +vint16m2_t vnmsub(vint16m2_t op0, int16_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m2_m))) +vint16m2_t vnmsub(vbool8_t op0, vint16m2_t op1, int16_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m4))) +vint16m4_t vnmsub(vint16m4_t op0, int16_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m4_m))) +vint16m4_t vnmsub(vbool4_t op0, vint16m4_t op1, int16_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m8))) +vint16m8_t vnmsub(vint16m8_t op0, int16_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m8_m))) +vint16m8_t vnmsub(vbool2_t op0, vint16m8_t op1, int16_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16mf2))) +vint16mf2_t vnmsub(vint16mf2_t op0, int16_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16mf2_m))) +vint16mf2_t vnmsub(vbool32_t op0, vint16mf2_t op1, int16_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16mf4))) +vint16mf4_t vnmsub(vint16mf4_t op0, int16_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16mf4_m))) +vint16mf4_t vnmsub(vbool64_t op0, vint16mf4_t op1, int16_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m1))) +vint32m1_t vnmsub(vint32m1_t op0, int32_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m1_m))) +vint32m1_t vnmsub(vbool32_t op0, vint32m1_t op1, int32_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m2))) +vint32m2_t vnmsub(vint32m2_t op0, int32_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m2_m))) +vint32m2_t vnmsub(vbool16_t op0, vint32m2_t op1, int32_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m4))) +vint32m4_t vnmsub(vint32m4_t op0, int32_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m4_m))) +vint32m4_t vnmsub(vbool8_t op0, vint32m4_t op1, int32_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m8))) +vint32m8_t vnmsub(vint32m8_t op0, int32_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m8_m))) +vint32m8_t vnmsub(vbool4_t op0, vint32m8_t op1, int32_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32mf2))) +vint32mf2_t vnmsub(vint32mf2_t op0, int32_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32mf2_m))) +vint32mf2_t vnmsub(vbool64_t op0, vint32mf2_t op1, int32_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m1))) +vint64m1_t vnmsub(vint64m1_t op0, int64_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m1_m))) +vint64m1_t vnmsub(vbool64_t op0, vint64m1_t op1, int64_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m2))) +vint64m2_t vnmsub(vint64m2_t op0, int64_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m2_m))) +vint64m2_t vnmsub(vbool32_t op0, vint64m2_t op1, int64_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m4))) +vint64m4_t vnmsub(vint64m4_t op0, int64_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m4_m))) +vint64m4_t vnmsub(vbool16_t op0, vint64m4_t op1, int64_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m8))) +vint64m8_t vnmsub(vint64m8_t op0, int64_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m8_m))) +vint64m8_t vnmsub(vbool8_t op0, vint64m8_t op1, int64_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m1))) +vuint8m1_t vnmsub(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m1_m))) +vuint8m1_t vnmsub(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m2))) +vuint8m2_t vnmsub(vuint8m2_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m2_m))) +vuint8m2_t vnmsub(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m4))) +vuint8m4_t vnmsub(vuint8m4_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m4_m))) +vuint8m4_t vnmsub(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m8))) +vuint8m8_t vnmsub(vuint8m8_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m8_m))) +vuint8m8_t vnmsub(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8mf2))) +vuint8mf2_t vnmsub(vuint8mf2_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8mf2_m))) +vuint8mf2_t vnmsub(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8mf4))) +vuint8mf4_t vnmsub(vuint8mf4_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8mf4_m))) +vuint8mf4_t vnmsub(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8mf8))) +vuint8mf8_t vnmsub(vuint8mf8_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8mf8_m))) +vuint8mf8_t vnmsub(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m1))) +vuint16m1_t vnmsub(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m1_m))) +vuint16m1_t vnmsub(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m2))) +vuint16m2_t vnmsub(vuint16m2_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m2_m))) +vuint16m2_t vnmsub(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m4))) +vuint16m4_t vnmsub(vuint16m4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m4_m))) +vuint16m4_t vnmsub(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m8))) +vuint16m8_t vnmsub(vuint16m8_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m8_m))) +vuint16m8_t vnmsub(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16mf2))) +vuint16mf2_t vnmsub(vuint16mf2_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16mf2_m))) +vuint16mf2_t vnmsub(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16mf4))) +vuint16mf4_t vnmsub(vuint16mf4_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16mf4_m))) +vuint16mf4_t vnmsub(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m1))) +vuint32m1_t vnmsub(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m1_m))) +vuint32m1_t vnmsub(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m2))) +vuint32m2_t vnmsub(vuint32m2_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m2_m))) +vuint32m2_t vnmsub(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m4))) +vuint32m4_t vnmsub(vuint32m4_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m4_m))) +vuint32m4_t vnmsub(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m8))) +vuint32m8_t vnmsub(vuint32m8_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m8_m))) +vuint32m8_t vnmsub(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32mf2))) +vuint32mf2_t vnmsub(vuint32mf2_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32mf2_m))) +vuint32mf2_t vnmsub(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m1))) +vuint64m1_t vnmsub(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m1_m))) +vuint64m1_t vnmsub(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m2))) +vuint64m2_t vnmsub(vuint64m2_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m2_m))) +vuint64m2_t vnmsub(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m4))) +vuint64m4_t vnmsub(vuint64m4_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m4_m))) +vuint64m4_t vnmsub(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m8))) +vuint64m8_t vnmsub(vuint64m8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m8_m))) +vuint64m8_t vnmsub(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m1))) +void vsuxei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m1_m))) +void vsuxei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m2))) +void vsuxei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m2_m))) +void vsuxei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m4))) +void vsuxei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m4_m))) +void vsuxei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m8))) +void vsuxei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m8_m))) +void vsuxei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32mf2))) +void vsuxei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32mf2_m))) +void vsuxei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m1))) +vuint8m1_t vnmsub(vuint8m1_t op0, uint8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m1_m))) +vuint8m1_t vnmsub(vbool8_t op0, vuint8m1_t op1, uint8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m2))) +vuint8m2_t vnmsub(vuint8m2_t op0, uint8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m2_m))) +vuint8m2_t vnmsub(vbool4_t op0, vuint8m2_t op1, uint8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m4))) +vuint8m4_t vnmsub(vuint8m4_t op0, uint8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m4_m))) +vuint8m4_t vnmsub(vbool2_t op0, vuint8m4_t op1, uint8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m8))) +vuint8m8_t vnmsub(vuint8m8_t op0, uint8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m8_m))) +vuint8m8_t vnmsub(vbool1_t op0, vuint8m8_t op1, uint8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8mf2))) +vuint8mf2_t vnmsub(vuint8mf2_t op0, uint8_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8mf2_m))) +vuint8mf2_t vnmsub(vbool16_t op0, vuint8mf2_t op1, uint8_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8mf4))) +vuint8mf4_t vnmsub(vuint8mf4_t op0, uint8_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8mf4_m))) +vuint8mf4_t vnmsub(vbool32_t op0, vuint8mf4_t op1, uint8_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8mf8))) +vuint8mf8_t vnmsub(vuint8mf8_t op0, uint8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8mf8_m))) +vuint8mf8_t vnmsub(vbool64_t op0, vuint8mf8_t op1, uint8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m1))) +vuint16m1_t vnmsub(vuint16m1_t op0, uint16_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m1_m))) +vuint16m1_t vnmsub(vbool16_t op0, vuint16m1_t op1, uint16_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m2))) +vuint16m2_t vnmsub(vuint16m2_t op0, uint16_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m2_m))) +vuint16m2_t vnmsub(vbool8_t op0, vuint16m2_t op1, uint16_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m4))) +vuint16m4_t vnmsub(vuint16m4_t op0, uint16_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m4_m))) +vuint16m4_t vnmsub(vbool4_t op0, vuint16m4_t op1, uint16_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m8))) +vuint16m8_t vnmsub(vuint16m8_t op0, uint16_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m8_m))) +vuint16m8_t vnmsub(vbool2_t op0, vuint16m8_t op1, uint16_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16mf2))) +vuint16mf2_t vnmsub(vuint16mf2_t op0, uint16_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16mf2_m))) +vuint16mf2_t vnmsub(vbool32_t op0, vuint16mf2_t op1, uint16_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16mf4))) +vuint16mf4_t vnmsub(vuint16mf4_t op0, uint16_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16mf4_m))) +vuint16mf4_t vnmsub(vbool64_t op0, vuint16mf4_t op1, uint16_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m1))) +vuint32m1_t vnmsub(vuint32m1_t op0, uint32_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m1_m))) +vuint32m1_t vnmsub(vbool32_t op0, vuint32m1_t op1, uint32_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m2))) +vuint32m2_t vnmsub(vuint32m2_t op0, uint32_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m2_m))) +vuint32m2_t vnmsub(vbool16_t op0, vuint32m2_t op1, uint32_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m4))) +vuint32m4_t vnmsub(vuint32m4_t op0, uint32_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m4_m))) +vuint32m4_t vnmsub(vbool8_t op0, vuint32m4_t op1, uint32_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m8))) +vuint32m8_t vnmsub(vuint32m8_t op0, uint32_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m8_m))) +vuint32m8_t vnmsub(vbool4_t op0, vuint32m8_t op1, uint32_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32mf2))) +vuint32mf2_t vnmsub(vuint32mf2_t op0, uint32_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32mf2_m))) +vuint32mf2_t vnmsub(vbool64_t op0, vuint32mf2_t op1, uint32_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m1))) +vuint64m1_t vnmsub(vuint64m1_t op0, uint64_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m1_m))) +vuint64m1_t vnmsub(vbool64_t op0, vuint64m1_t op1, uint64_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m2))) +vuint64m2_t vnmsub(vuint64m2_t op0, uint64_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m2_m))) +vuint64m2_t vnmsub(vbool32_t op0, vuint64m2_t op1, uint64_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m4))) +vuint64m4_t vnmsub(vuint64m4_t op0, uint64_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m4_m))) +vuint64m4_t vnmsub(vbool16_t op0, vuint64m4_t op1, uint64_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m8))) +vuint64m8_t vnmsub(vuint64m8_t op0, uint64_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m8_m))) +vuint64m8_t vnmsub(vbool8_t op0, vuint64m8_t op1, uint64_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16mf4))) +vuint16mf4_t vwmaccu(vuint16mf4_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16mf4_m))) +vuint16mf4_t vwmaccu(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16mf2))) +vuint16mf2_t vwmaccu(vuint16mf2_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16mf2_m))) +vuint16mf2_t vwmaccu(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m1))) +vuint16m1_t vwmaccu(vuint16m1_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m1_m))) +vuint16m1_t vwmaccu(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m2))) +vuint16m2_t vwmaccu(vuint16m2_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m2_m))) +vuint16m2_t vwmaccu(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m4))) +vuint16m4_t vwmaccu(vuint16m4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m4_m))) +vuint16m4_t vwmaccu(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m8))) +vuint16m8_t vwmaccu(vuint16m8_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m8_m))) +vuint16m8_t vwmaccu(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32mf2))) +vuint32mf2_t vwmaccu(vuint32mf2_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32mf2_m))) +vuint32mf2_t vwmaccu(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m1))) +vuint32m1_t vwmaccu(vuint32m1_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m1_m))) +vuint32m1_t vwmaccu(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m2))) +vuint32m2_t vwmaccu(vuint32m2_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m2_m))) +vuint32m2_t vwmaccu(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m4))) +vuint32m4_t vwmaccu(vuint32m4_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m4_m))) +vuint32m4_t vwmaccu(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m8))) +vuint32m8_t vwmaccu(vuint32m8_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m8_m))) +vuint32m8_t vwmaccu(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m1))) +vuint64m1_t vwmaccu(vuint64m1_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m1_m))) +vuint64m1_t vwmaccu(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m2))) +vuint64m2_t vwmaccu(vuint64m2_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m2_m))) +vuint64m2_t vwmaccu(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m4))) +vuint64m4_t vwmaccu(vuint64m4_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m4_m))) +vuint64m4_t vwmaccu(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m8))) +vuint64m8_t vwmaccu(vuint64m8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m8_m))) +vuint64m8_t vwmaccu(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16mf4))) +vuint16mf4_t vwmaccu(vuint16mf4_t op0, uint8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16mf4_m))) +vuint16mf4_t vwmaccu(vbool64_t op0, vuint16mf4_t op1, uint8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16mf2))) +vuint16mf2_t vwmaccu(vuint16mf2_t op0, uint8_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16mf2_m))) +vuint16mf2_t vwmaccu(vbool32_t op0, vuint16mf2_t op1, uint8_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m1))) +vuint16m1_t vwmaccu(vuint16m1_t op0, uint8_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m1_m))) +vuint16m1_t vwmaccu(vbool16_t op0, vuint16m1_t op1, uint8_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m2))) +vuint16m2_t vwmaccu(vuint16m2_t op0, uint8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m2_m))) +vuint16m2_t vwmaccu(vbool8_t op0, vuint16m2_t op1, uint8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m4))) +vuint16m4_t vwmaccu(vuint16m4_t op0, uint8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m4_m))) +vuint16m4_t vwmaccu(vbool4_t op0, vuint16m4_t op1, uint8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m8))) +vuint16m8_t vwmaccu(vuint16m8_t op0, uint8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m8_m))) +vuint16m8_t vwmaccu(vbool2_t op0, vuint16m8_t op1, uint8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32mf2))) +vuint32mf2_t vwmaccu(vuint32mf2_t op0, uint16_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32mf2_m))) +vuint32mf2_t vwmaccu(vbool64_t op0, vuint32mf2_t op1, uint16_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m1))) +vuint32m1_t vwmaccu(vuint32m1_t op0, uint16_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m1_m))) +vuint32m1_t vwmaccu(vbool32_t op0, vuint32m1_t op1, uint16_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m2))) +vuint32m2_t vwmaccu(vuint32m2_t op0, uint16_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m2_m))) +vuint32m2_t vwmaccu(vbool16_t op0, vuint32m2_t op1, uint16_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m4))) +vuint32m4_t vwmaccu(vuint32m4_t op0, uint16_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m4_m))) +vuint32m4_t vwmaccu(vbool8_t op0, vuint32m4_t op1, uint16_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m8))) +vuint32m8_t vwmaccu(vuint32m8_t op0, uint16_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m8_m))) +vuint32m8_t vwmaccu(vbool4_t op0, vuint32m8_t op1, uint16_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m1))) +vuint64m1_t vwmaccu(vuint64m1_t op0, uint32_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m1_m))) +vuint64m1_t vwmaccu(vbool64_t op0, vuint64m1_t op1, uint32_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m2))) +vuint64m2_t vwmaccu(vuint64m2_t op0, uint32_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m2_m))) +vuint64m2_t vwmaccu(vbool32_t op0, vuint64m2_t op1, uint32_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m4))) +vuint64m4_t vwmaccu(vuint64m4_t op0, uint32_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m4_m))) +vuint64m4_t vwmaccu(vbool16_t op0, vuint64m4_t op1, uint32_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m8))) +vuint64m8_t vwmaccu(vuint64m8_t op0, uint32_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m8_m))) +vuint64m8_t vwmaccu(vbool8_t op0, vuint64m8_t op1, uint32_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16mf4))) +vint16mf4_t vwmacc(vint16mf4_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16mf4_m))) +vint16mf4_t vwmacc(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16mf2))) +vint16mf2_t vwmacc(vint16mf2_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16mf2_m))) +vint16mf2_t vwmacc(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m1))) +vint16m1_t vwmacc(vint16m1_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m1_m))) +vint16m1_t vwmacc(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m2))) +vint16m2_t vwmacc(vint16m2_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m2_m))) +vint16m2_t vwmacc(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m4))) +vint16m4_t vwmacc(vint16m4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m4_m))) +vint16m4_t vwmacc(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m8))) +vint16m8_t vwmacc(vint16m8_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m8_m))) +vint16m8_t vwmacc(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32mf2))) +vint32mf2_t vwmacc(vint32mf2_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32mf2_m))) +vint32mf2_t vwmacc(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m1))) +vint32m1_t vwmacc(vint32m1_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m1_m))) +vint32m1_t vwmacc(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m2))) +vint32m2_t vwmacc(vint32m2_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m2_m))) +vint32m2_t vwmacc(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m4))) +vint32m4_t vwmacc(vint32m4_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m4_m))) +vint32m4_t vwmacc(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m8))) +vint32m8_t vwmacc(vint32m8_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m8_m))) +vint32m8_t vwmacc(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m1))) +vint64m1_t vwmacc(vint64m1_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m1_m))) +vint64m1_t vwmacc(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m2))) +vint64m2_t vwmacc(vint64m2_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m2_m))) +vint64m2_t vwmacc(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m4))) +vint64m4_t vwmacc(vint64m4_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m4_m))) +vint64m4_t vwmacc(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m8))) +vint64m8_t vwmacc(vint64m8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m8_m))) +vint64m8_t vwmacc(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16mf4))) +vint16mf4_t vwmacc(vint16mf4_t op0, int8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16mf4_m))) +vint16mf4_t vwmacc(vbool64_t op0, vint16mf4_t op1, int8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16mf2))) +vint16mf2_t vwmacc(vint16mf2_t op0, int8_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16mf2_m))) +vint16mf2_t vwmacc(vbool32_t op0, vint16mf2_t op1, int8_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m1))) +vint16m1_t vwmacc(vint16m1_t op0, int8_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m1_m))) +vint16m1_t vwmacc(vbool16_t op0, vint16m1_t op1, int8_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m2))) +vint16m2_t vwmacc(vint16m2_t op0, int8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m2_m))) +vint16m2_t vwmacc(vbool8_t op0, vint16m2_t op1, int8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m4))) +vint16m4_t vwmacc(vint16m4_t op0, int8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m4_m))) +vint16m4_t vwmacc(vbool4_t op0, vint16m4_t op1, int8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m8))) +vint16m8_t vwmacc(vint16m8_t op0, int8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m8_m))) +vint16m8_t vwmacc(vbool2_t op0, vint16m8_t op1, int8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32mf2))) +vint32mf2_t vwmacc(vint32mf2_t op0, int16_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32mf2_m))) +vint32mf2_t vwmacc(vbool64_t op0, vint32mf2_t op1, int16_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m1))) +vint32m1_t vwmacc(vint32m1_t op0, int16_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m1_m))) +vint32m1_t vwmacc(vbool32_t op0, vint32m1_t op1, int16_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m2))) +vint32m2_t vwmacc(vint32m2_t op0, int16_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m2_m))) +vint32m2_t vwmacc(vbool16_t op0, vint32m2_t op1, int16_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m4))) +vint32m4_t vwmacc(vint32m4_t op0, int16_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m4_m))) +vint32m4_t vwmacc(vbool8_t op0, vint32m4_t op1, int16_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m8))) +vint32m8_t vwmacc(vint32m8_t op0, int16_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m8_m))) +vint32m8_t vwmacc(vbool4_t op0, vint32m8_t op1, int16_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m1))) +vint64m1_t vwmacc(vint64m1_t op0, int32_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m1_m))) +vint64m1_t vwmacc(vbool64_t op0, vint64m1_t op1, int32_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m2))) +vint64m2_t vwmacc(vint64m2_t op0, int32_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m2_m))) +vint64m2_t vwmacc(vbool32_t op0, vint64m2_t op1, int32_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m4))) +vint64m4_t vwmacc(vint64m4_t op0, int32_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m4_m))) +vint64m4_t vwmacc(vbool16_t op0, vint64m4_t op1, int32_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m8))) +vint64m8_t vwmacc(vint64m8_t op0, int32_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m8_m))) +vint64m8_t vwmacc(vbool8_t op0, vint64m8_t op1, int32_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16mf4))) +vint16mf4_t vwmaccsu(vint16mf4_t op0, vint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16mf4_m))) +vint16mf4_t vwmaccsu(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16mf2))) +vint16mf2_t vwmaccsu(vint16mf2_t op0, vint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16mf2_m))) +vint16mf2_t vwmaccsu(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m1))) +vint16m1_t vwmaccsu(vint16m1_t op0, vint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m1_m))) +vint16m1_t vwmaccsu(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m2))) +vint16m2_t vwmaccsu(vint16m2_t op0, vint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m2_m))) +vint16m2_t vwmaccsu(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m4))) +vint16m4_t vwmaccsu(vint16m4_t op0, vint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m4_m))) +vint16m4_t vwmaccsu(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m8))) +vint16m8_t vwmaccsu(vint16m8_t op0, vint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m8_m))) +vint16m8_t vwmaccsu(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32mf2))) +vint32mf2_t vwmaccsu(vint32mf2_t op0, vint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32mf2_m))) +vint32mf2_t vwmaccsu(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m1))) +vint32m1_t vwmaccsu(vint32m1_t op0, vint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m1_m))) +vint32m1_t vwmaccsu(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m2))) +vint32m2_t vwmaccsu(vint32m2_t op0, vint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m2_m))) +vint32m2_t vwmaccsu(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m4))) +vint32m4_t vwmaccsu(vint32m4_t op0, vint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m4_m))) +vint32m4_t vwmaccsu(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m8))) +vint32m8_t vwmaccsu(vint32m8_t op0, vint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m8_m))) +vint32m8_t vwmaccsu(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m1))) +vint64m1_t vwmaccsu(vint64m1_t op0, vint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m1_m))) +vint64m1_t vwmaccsu(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m2))) +vint64m2_t vwmaccsu(vint64m2_t op0, vint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m2_m))) +vint64m2_t vwmaccsu(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m4))) +vint64m4_t vwmaccsu(vint64m4_t op0, vint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m4_m))) +vint64m4_t vwmaccsu(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m8))) +vint64m8_t vwmaccsu(vint64m8_t op0, vint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m8_m))) +vint64m8_t vwmaccsu(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16mf4))) +vint16mf4_t vwmaccsu(vint16mf4_t op0, int8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16mf4_m))) +vint16mf4_t vwmaccsu(vbool64_t op0, vint16mf4_t op1, int8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16mf2))) +vint16mf2_t vwmaccsu(vint16mf2_t op0, int8_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16mf2_m))) +vint16mf2_t vwmaccsu(vbool32_t op0, vint16mf2_t op1, int8_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m1))) +vint16m1_t vwmaccsu(vint16m1_t op0, int8_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m1_m))) +vint16m1_t vwmaccsu(vbool16_t op0, vint16m1_t op1, int8_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m2))) +vint16m2_t vwmaccsu(vint16m2_t op0, int8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m2_m))) +vint16m2_t vwmaccsu(vbool8_t op0, vint16m2_t op1, int8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m4))) +vint16m4_t vwmaccsu(vint16m4_t op0, int8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m4_m))) +vint16m4_t vwmaccsu(vbool4_t op0, vint16m4_t op1, int8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m8))) +vint16m8_t vwmaccsu(vint16m8_t op0, int8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m8_m))) +vint16m8_t vwmaccsu(vbool2_t op0, vint16m8_t op1, int8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32mf2))) +vint32mf2_t vwmaccsu(vint32mf2_t op0, int16_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32mf2_m))) +vint32mf2_t vwmaccsu(vbool64_t op0, vint32mf2_t op1, int16_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m1))) +vint32m1_t vwmaccsu(vint32m1_t op0, int16_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m1_m))) +vint32m1_t vwmaccsu(vbool32_t op0, vint32m1_t op1, int16_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m2))) +vint32m2_t vwmaccsu(vint32m2_t op0, int16_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m2_m))) +vint32m2_t vwmaccsu(vbool16_t op0, vint32m2_t op1, int16_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m4))) +vint32m4_t vwmaccsu(vint32m4_t op0, int16_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m4_m))) +vint32m4_t vwmaccsu(vbool8_t op0, vint32m4_t op1, int16_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m8))) +vint32m8_t vwmaccsu(vint32m8_t op0, int16_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m8_m))) +vint32m8_t vwmaccsu(vbool4_t op0, vint32m8_t op1, int16_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m1))) +vint64m1_t vwmaccsu(vint64m1_t op0, int32_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m1_m))) +vint64m1_t vwmaccsu(vbool64_t op0, vint64m1_t op1, int32_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m2))) +vint64m2_t vwmaccsu(vint64m2_t op0, int32_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m2_m))) +vint64m2_t vwmaccsu(vbool32_t op0, vint64m2_t op1, int32_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m4))) +vint64m4_t vwmaccsu(vint64m4_t op0, int32_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m4_m))) +vint64m4_t vwmaccsu(vbool16_t op0, vint64m4_t op1, int32_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m8))) +vint64m8_t vwmaccsu(vint64m8_t op0, int32_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m8_m))) +vint64m8_t vwmaccsu(vbool8_t op0, vint64m8_t op1, int32_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16mf4))) +vint16mf4_t vwmaccus(vint16mf4_t op0, uint8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16mf4_m))) +vint16mf4_t vwmaccus(vbool64_t op0, vint16mf4_t op1, uint8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16mf2))) +vint16mf2_t vwmaccus(vint16mf2_t op0, uint8_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16mf2_m))) +vint16mf2_t vwmaccus(vbool32_t op0, vint16mf2_t op1, uint8_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m1))) +vint16m1_t vwmaccus(vint16m1_t op0, uint8_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m1_m))) +vint16m1_t vwmaccus(vbool16_t op0, vint16m1_t op1, uint8_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m2))) +vint16m2_t vwmaccus(vint16m2_t op0, uint8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m2_m))) +vint16m2_t vwmaccus(vbool8_t op0, vint16m2_t op1, uint8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m4))) +vint16m4_t vwmaccus(vint16m4_t op0, uint8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m4_m))) +vint16m4_t vwmaccus(vbool4_t op0, vint16m4_t op1, uint8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m8))) +vint16m8_t vwmaccus(vint16m8_t op0, uint8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m8_m))) +vint16m8_t vwmaccus(vbool2_t op0, vint16m8_t op1, uint8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32mf2))) +vint32mf2_t vwmaccus(vint32mf2_t op0, uint16_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32mf2_m))) +vint32mf2_t vwmaccus(vbool64_t op0, vint32mf2_t op1, uint16_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m1))) +vint32m1_t vwmaccus(vint32m1_t op0, uint16_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m1_m))) +vint32m1_t vwmaccus(vbool32_t op0, vint32m1_t op1, uint16_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m2))) +vint32m2_t vwmaccus(vint32m2_t op0, uint16_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m2_m))) +vint32m2_t vwmaccus(vbool16_t op0, vint32m2_t op1, uint16_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m4))) +vint32m4_t vwmaccus(vint32m4_t op0, uint16_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m4_m))) +vint32m4_t vwmaccus(vbool8_t op0, vint32m4_t op1, uint16_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m8))) +vint32m8_t vwmaccus(vint32m8_t op0, uint16_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m8_m))) +vint32m8_t vwmaccus(vbool4_t op0, vint32m8_t op1, uint16_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m1))) +vint64m1_t vwmaccus(vint64m1_t op0, uint32_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m1_m))) +vint64m1_t vwmaccus(vbool64_t op0, vint64m1_t op1, uint32_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m2))) +vint64m2_t vwmaccus(vint64m2_t op0, uint32_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m2_m))) +vint64m2_t vwmaccus(vbool32_t op0, vint64m2_t op1, uint32_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m4))) +vint64m4_t vwmaccus(vint64m4_t op0, uint32_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m4_m))) +vint64m4_t vwmaccus(vbool16_t op0, vint64m4_t op1, uint32_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m8))) +vint64m8_t vwmaccus(vint64m8_t op0, uint32_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m8_m))) +vint64m8_t vwmaccus(vbool8_t op0, vint64m8_t op1, uint32_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8m1))) +vint8m1_t vmerge(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8m2))) +vint8m2_t vmerge(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8m4))) +vint8m4_t vmerge(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8m8))) +vint8m8_t vmerge(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8mf2))) +vint8mf2_t vmerge(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8mf4))) +vint8mf4_t vmerge(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8mf8))) +vint8mf8_t vmerge(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i16m1))) +vint16m1_t vmerge(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i16m2))) +vint16m2_t vmerge(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i16m4))) +vint16m4_t vmerge(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i16m8))) +vint16m8_t vmerge(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i16mf2))) +vint16mf2_t vmerge(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i16mf4))) +vint16mf4_t vmerge(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i32m1))) +vint32m1_t vmerge(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i32m2))) +vint32m2_t vmerge(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i32m4))) +vint32m4_t vmerge(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i32m8))) +vint32m8_t vmerge(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i32mf2))) +vint32mf2_t vmerge(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i64m1))) +vint64m1_t vmerge(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i64m2))) +vint64m2_t vmerge(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i64m4))) +vint64m4_t vmerge(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i64m8))) +vint64m8_t vmerge(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8m1))) +vint8m1_t vmerge(vbool8_t op0, vint8m1_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8m2))) +vint8m2_t vmerge(vbool4_t op0, vint8m2_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8m4))) +vint8m4_t vmerge(vbool2_t op0, vint8m4_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8m8))) +vint8m8_t vmerge(vbool1_t op0, vint8m8_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8mf2))) +vint8mf2_t vmerge(vbool16_t op0, vint8mf2_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8mf4))) +vint8mf4_t vmerge(vbool32_t op0, vint8mf4_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8mf8))) +vint8mf8_t vmerge(vbool64_t op0, vint8mf8_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i16m1))) +vint16m1_t vmerge(vbool16_t op0, vint16m1_t op1, int16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i16m2))) +vint16m2_t vmerge(vbool8_t op0, vint16m2_t op1, int16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i16m4))) +vint16m4_t vmerge(vbool4_t op0, vint16m4_t op1, int16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i16m8))) +vint16m8_t vmerge(vbool2_t op0, vint16m8_t op1, int16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i16mf2))) +vint16mf2_t vmerge(vbool32_t op0, vint16mf2_t op1, int16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i16mf4))) +vint16mf4_t vmerge(vbool64_t op0, vint16mf4_t op1, int16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i32m1))) +vint32m1_t vmerge(vbool32_t op0, vint32m1_t op1, int32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i32m2))) +vint32m2_t vmerge(vbool16_t op0, vint32m2_t op1, int32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i32m4))) +vint32m4_t vmerge(vbool8_t op0, vint32m4_t op1, int32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i32m8))) +vint32m8_t vmerge(vbool4_t op0, vint32m8_t op1, int32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i32mf2))) +vint32mf2_t vmerge(vbool64_t op0, vint32mf2_t op1, int32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i64m1))) +vint64m1_t vmerge(vbool64_t op0, vint64m1_t op1, int64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i64m2))) +vint64m2_t vmerge(vbool32_t op0, vint64m2_t op1, int64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i64m4))) +vint64m4_t vmerge(vbool16_t op0, vint64m4_t op1, int64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i64m8))) +vint64m8_t vmerge(vbool8_t op0, vint64m8_t op1, int64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m1))) +void vsuxei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m1_m))) +void vsuxei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m2))) +void vsuxei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m2_m))) +void vsuxei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m4))) +void vsuxei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m4_m))) +void vsuxei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m8))) +void vsuxei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m8_m))) +void vsuxei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32mf2))) +void vsuxei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32mf2_m))) +void vsuxei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8m1))) +vuint8m1_t vmerge(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8m2))) +vuint8m2_t vmerge(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8m4))) +vuint8m4_t vmerge(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8m8))) +vuint8m8_t vmerge(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8mf2))) +vuint8mf2_t vmerge(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8mf4))) +vuint8mf4_t vmerge(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8mf8))) +vuint8mf8_t vmerge(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u16m1))) +vuint16m1_t vmerge(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u16m2))) +vuint16m2_t vmerge(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u16m4))) +vuint16m4_t vmerge(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u16m8))) +vuint16m8_t vmerge(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u16mf2))) +vuint16mf2_t vmerge(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u16mf4))) +vuint16mf4_t vmerge(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u32m1))) +vuint32m1_t vmerge(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u32m2))) +vuint32m2_t vmerge(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u32m4))) +vuint32m4_t vmerge(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u32m8))) +vuint32m8_t vmerge(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u32mf2))) +vuint32mf2_t vmerge(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u64m1))) +vuint64m1_t vmerge(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u64m2))) +vuint64m2_t vmerge(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u64m4))) +vuint64m4_t vmerge(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u64m8))) +vuint64m8_t vmerge(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8m1))) +vuint8m1_t vmerge(vbool8_t op0, vuint8m1_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8m2))) +vuint8m2_t vmerge(vbool4_t op0, vuint8m2_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8m4))) +vuint8m4_t vmerge(vbool2_t op0, vuint8m4_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8m8))) +vuint8m8_t vmerge(vbool1_t op0, vuint8m8_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8mf2))) +vuint8mf2_t vmerge(vbool16_t op0, vuint8mf2_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8mf4))) +vuint8mf4_t vmerge(vbool32_t op0, vuint8mf4_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8mf8))) +vuint8mf8_t vmerge(vbool64_t op0, vuint8mf8_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u16m1))) +vuint16m1_t vmerge(vbool16_t op0, vuint16m1_t op1, uint16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u16m2))) +vuint16m2_t vmerge(vbool8_t op0, vuint16m2_t op1, uint16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u16m4))) +vuint16m4_t vmerge(vbool4_t op0, vuint16m4_t op1, uint16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u16m8))) +vuint16m8_t vmerge(vbool2_t op0, vuint16m8_t op1, uint16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u16mf2))) +vuint16mf2_t vmerge(vbool32_t op0, vuint16mf2_t op1, uint16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u16mf4))) +vuint16mf4_t vmerge(vbool64_t op0, vuint16mf4_t op1, uint16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u32m1))) +vuint32m1_t vmerge(vbool32_t op0, vuint32m1_t op1, uint32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u32m2))) +vuint32m2_t vmerge(vbool16_t op0, vuint32m2_t op1, uint32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u32m4))) +vuint32m4_t vmerge(vbool8_t op0, vuint32m4_t op1, uint32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u32m8))) +vuint32m8_t vmerge(vbool4_t op0, vuint32m8_t op1, uint32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u32mf2))) +vuint32mf2_t vmerge(vbool64_t op0, vuint32mf2_t op1, uint32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u64m1))) +vuint64m1_t vmerge(vbool64_t op0, vuint64m1_t op1, uint64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u64m2))) +vuint64m2_t vmerge(vbool32_t op0, vuint64m2_t op1, uint64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u64m4))) +vuint64m4_t vmerge(vbool16_t op0, vuint64m4_t op1, uint64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u64m8))) +vuint64m8_t vmerge(vbool8_t op0, vuint64m8_t op1, uint64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8m1))) +vuint8m1_t vmv_v(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8m2))) +vuint8m2_t vmv_v(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8m4))) +vuint8m4_t vmv_v(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8m8))) +vuint8m8_t vmv_v(vuint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8mf2))) +vuint8mf2_t vmv_v(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8mf4))) +vuint8mf4_t vmv_v(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8mf8))) +vuint8mf8_t vmv_v(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u16m1))) +vuint16m1_t vmv_v(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u16m2))) +vuint16m2_t vmv_v(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u16m4))) +vuint16m4_t vmv_v(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u16m8))) +vuint16m8_t vmv_v(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u16mf2))) +vuint16mf2_t vmv_v(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u16mf4))) +vuint16mf4_t vmv_v(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u32m1))) +vuint32m1_t vmv_v(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u32m2))) +vuint32m2_t vmv_v(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u32m4))) +vuint32m4_t vmv_v(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u32m8))) +vuint32m8_t vmv_v(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u32mf2))) +vuint32mf2_t vmv_v(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u64m1))) +vuint64m1_t vmv_v(vuint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u64m2))) +vuint64m2_t vmv_v(vuint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u64m4))) +vuint64m4_t vmv_v(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u64m8))) +vuint64m8_t vmv_v(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8m1))) +vint8m1_t vmv_v(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8m2))) +vint8m2_t vmv_v(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8m4))) +vint8m4_t vmv_v(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8m8))) +vint8m8_t vmv_v(vint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8mf2))) +vint8mf2_t vmv_v(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8mf4))) +vint8mf4_t vmv_v(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8mf8))) +vint8mf8_t vmv_v(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i16m1))) +vint16m1_t vmv_v(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i16m2))) +vint16m2_t vmv_v(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i16m4))) +vint16m4_t vmv_v(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i16m8))) +vint16m8_t vmv_v(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i16mf2))) +vint16mf2_t vmv_v(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i16mf4))) +vint16mf4_t vmv_v(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i32m1))) +vint32m1_t vmv_v(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i32m2))) +vint32m2_t vmv_v(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i32m4))) +vint32m4_t vmv_v(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i32m8))) +vint32m8_t vmv_v(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i32mf2))) +vint32mf2_t vmv_v(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i64m1))) +vint64m1_t vmv_v(vint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i64m2))) +vint64m2_t vmv_v(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i64m4))) +vint64m4_t vmv_v(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i64m8))) +vint64m8_t vmv_v(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m1))) +vuint8m1_t vsaddu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m1_m))) +vuint8m1_t vsaddu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m2))) +vuint8m2_t vsaddu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m2_m))) +vuint8m2_t vsaddu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m4))) +vuint8m4_t vsaddu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m4_m))) +vuint8m4_t vsaddu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m8))) +vuint8m8_t vsaddu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m8_m))) +vuint8m8_t vsaddu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8mf2))) +vuint8mf2_t vsaddu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8mf2_m))) +vuint8mf2_t vsaddu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8mf4))) +vuint8mf4_t vsaddu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8mf4_m))) +vuint8mf4_t vsaddu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8mf8))) +vuint8mf8_t vsaddu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8mf8_m))) +vuint8mf8_t vsaddu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m1))) +vuint16m1_t vsaddu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m1_m))) +vuint16m1_t vsaddu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m2))) +vuint16m2_t vsaddu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m2_m))) +vuint16m2_t vsaddu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m4))) +vuint16m4_t vsaddu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m4_m))) +vuint16m4_t vsaddu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m8))) +vuint16m8_t vsaddu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m8_m))) +vuint16m8_t vsaddu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16mf2))) +vuint16mf2_t vsaddu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16mf2_m))) +vuint16mf2_t vsaddu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16mf4))) +vuint16mf4_t vsaddu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16mf4_m))) +vuint16mf4_t vsaddu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m1))) +vuint32m1_t vsaddu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m1_m))) +vuint32m1_t vsaddu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m2))) +vuint32m2_t vsaddu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m2_m))) +vuint32m2_t vsaddu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m4))) +vuint32m4_t vsaddu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m4_m))) +vuint32m4_t vsaddu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m8))) +vuint32m8_t vsaddu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m8_m))) +vuint32m8_t vsaddu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32mf2))) +vuint32mf2_t vsaddu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32mf2_m))) +vuint32mf2_t vsaddu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m1))) +vuint64m1_t vsaddu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m1_m))) +vuint64m1_t vsaddu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m2))) +vuint64m2_t vsaddu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m2_m))) +vuint64m2_t vsaddu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m4))) +vuint64m4_t vsaddu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m4_m))) +vuint64m4_t vsaddu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m8))) +vuint64m8_t vsaddu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m8_m))) +vuint64m8_t vsaddu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m1))) +vuint8m1_t vsaddu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m1_m))) +vuint8m1_t vsaddu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m2))) +vuint8m2_t vsaddu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m2_m))) +vuint8m2_t vsaddu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m4))) +vuint8m4_t vsaddu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m4_m))) +vuint8m4_t vsaddu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m8))) +vuint8m8_t vsaddu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m8_m))) +vuint8m8_t vsaddu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8mf2))) +vuint8mf2_t vsaddu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8mf2_m))) +vuint8mf2_t vsaddu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8mf4))) +vuint8mf4_t vsaddu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8mf4_m))) +vuint8mf4_t vsaddu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8mf8))) +vuint8mf8_t vsaddu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8mf8_m))) +vuint8mf8_t vsaddu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m1))) +vuint16m1_t vsaddu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m1_m))) +vuint16m1_t vsaddu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m2))) +vuint16m2_t vsaddu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m2_m))) +vuint16m2_t vsaddu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m4))) +vuint16m4_t vsaddu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m4_m))) +vuint16m4_t vsaddu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m8))) +vuint16m8_t vsaddu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m8_m))) +vuint16m8_t vsaddu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16mf2))) +vuint16mf2_t vsaddu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16mf2_m))) +vuint16mf2_t vsaddu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16mf4))) +vuint16mf4_t vsaddu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16mf4_m))) +vuint16mf4_t vsaddu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m1))) +vuint32m1_t vsaddu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m1_m))) +vuint32m1_t vsaddu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m2))) +vuint32m2_t vsaddu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m2_m))) +vuint32m2_t vsaddu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m4))) +vuint32m4_t vsaddu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m4_m))) +vuint32m4_t vsaddu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m8))) +vuint32m8_t vsaddu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m8_m))) +vuint32m8_t vsaddu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32mf2))) +vuint32mf2_t vsaddu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32mf2_m))) +vuint32mf2_t vsaddu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m1))) +vuint64m1_t vsaddu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m1_m))) +vuint64m1_t vsaddu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m2))) +vuint64m2_t vsaddu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m2_m))) +vuint64m2_t vsaddu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m4))) +vuint64m4_t vsaddu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m4_m))) +vuint64m4_t vsaddu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m8))) +vuint64m8_t vsaddu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m8_m))) +vuint64m8_t vsaddu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m1))) +vint8m1_t vsadd(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m1_m))) +vint8m1_t vsadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m2))) +vint8m2_t vsadd(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m2_m))) +vint8m2_t vsadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m4))) +vint8m4_t vsadd(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m4_m))) +vint8m4_t vsadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m8))) +vint8m8_t vsadd(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m8_m))) +vint8m8_t vsadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8mf2))) +vint8mf2_t vsadd(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8mf2_m))) +vint8mf2_t vsadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8mf4))) +vint8mf4_t vsadd(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8mf4_m))) +vint8mf4_t vsadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8mf8))) +vint8mf8_t vsadd(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8mf8_m))) +vint8mf8_t vsadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m1))) +vint16m1_t vsadd(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m1_m))) +vint16m1_t vsadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m2))) +vint16m2_t vsadd(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m2_m))) +vint16m2_t vsadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m4))) +vint16m4_t vsadd(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m4_m))) +vint16m4_t vsadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m8))) +vint16m8_t vsadd(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m8_m))) +vint16m8_t vsadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16mf2))) +vint16mf2_t vsadd(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16mf2_m))) +vint16mf2_t vsadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16mf4))) +vint16mf4_t vsadd(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16mf4_m))) +vint16mf4_t vsadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m1))) +vint32m1_t vsadd(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m1_m))) +vint32m1_t vsadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m2))) +vint32m2_t vsadd(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m2_m))) +vint32m2_t vsadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m4))) +vint32m4_t vsadd(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m4_m))) +vint32m4_t vsadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m8))) +vint32m8_t vsadd(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m8_m))) +vint32m8_t vsadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32mf2))) +vint32mf2_t vsadd(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32mf2_m))) +vint32mf2_t vsadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m1))) +vint64m1_t vsadd(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m1_m))) +vint64m1_t vsadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m2))) +vint64m2_t vsadd(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m2_m))) +vint64m2_t vsadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m4))) +vint64m4_t vsadd(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m4_m))) +vint64m4_t vsadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m8))) +vint64m8_t vsadd(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m8_m))) +vint64m8_t vsadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m1))) +vint8m1_t vsadd(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m1_m))) +vint8m1_t vsadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m2))) +vint8m2_t vsadd(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m2_m))) +vint8m2_t vsadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m4))) +vint8m4_t vsadd(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m4_m))) +vint8m4_t vsadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m8))) +vint8m8_t vsadd(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m8_m))) +vint8m8_t vsadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8mf2))) +vint8mf2_t vsadd(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8mf2_m))) +vint8mf2_t vsadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8mf4))) +vint8mf4_t vsadd(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8mf4_m))) +vint8mf4_t vsadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8mf8))) +vint8mf8_t vsadd(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8mf8_m))) +vint8mf8_t vsadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m1))) +vint16m1_t vsadd(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m1_m))) +vint16m1_t vsadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m2))) +vint16m2_t vsadd(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m2_m))) +vint16m2_t vsadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m4))) +vint16m4_t vsadd(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m4_m))) +vint16m4_t vsadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m8))) +vint16m8_t vsadd(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m8_m))) +vint16m8_t vsadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16mf2))) +vint16mf2_t vsadd(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16mf2_m))) +vint16mf2_t vsadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16mf4))) +vint16mf4_t vsadd(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16mf4_m))) +vint16mf4_t vsadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m1))) +vint32m1_t vsadd(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m1_m))) +vint32m1_t vsadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m2))) +vint32m2_t vsadd(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m2_m))) +vint32m2_t vsadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m4))) +vint32m4_t vsadd(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m4_m))) +vint32m4_t vsadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m8))) +vint32m8_t vsadd(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m8_m))) +vint32m8_t vsadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32mf2))) +vint32mf2_t vsadd(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32mf2_m))) +vint32mf2_t vsadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m1))) +vint64m1_t vsadd(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m1_m))) +vint64m1_t vsadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m2))) +vint64m2_t vsadd(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m2_m))) +vint64m2_t vsadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m4))) +vint64m4_t vsadd(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m4_m))) +vint64m4_t vsadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m8))) +vint64m8_t vsadd(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m8_m))) +vint64m8_t vsadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m1))) +void vsuxei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m1_m))) +void vsuxei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m2))) +void vsuxei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m2_m))) +void vsuxei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m4))) +void vsuxei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m4_m))) +void vsuxei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m8))) +void vsuxei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m8_m))) +void vsuxei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32mf2))) +void vsuxei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32mf2_m))) +void vsuxei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m1))) +vuint8m1_t vssubu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m1_m))) +vuint8m1_t vssubu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m2))) +vuint8m2_t vssubu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m2_m))) +vuint8m2_t vssubu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m4))) +vuint8m4_t vssubu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m4_m))) +vuint8m4_t vssubu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m8))) +vuint8m8_t vssubu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m8_m))) +vuint8m8_t vssubu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8mf2))) +vuint8mf2_t vssubu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8mf2_m))) +vuint8mf2_t vssubu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8mf4))) +vuint8mf4_t vssubu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8mf4_m))) +vuint8mf4_t vssubu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8mf8))) +vuint8mf8_t vssubu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8mf8_m))) +vuint8mf8_t vssubu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m1))) +vuint16m1_t vssubu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m1_m))) +vuint16m1_t vssubu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m2))) +vuint16m2_t vssubu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m2_m))) +vuint16m2_t vssubu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m4))) +vuint16m4_t vssubu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m4_m))) +vuint16m4_t vssubu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m8))) +vuint16m8_t vssubu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m8_m))) +vuint16m8_t vssubu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16mf2))) +vuint16mf2_t vssubu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16mf2_m))) +vuint16mf2_t vssubu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16mf4))) +vuint16mf4_t vssubu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16mf4_m))) +vuint16mf4_t vssubu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m1))) +vuint32m1_t vssubu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m1_m))) +vuint32m1_t vssubu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m2))) +vuint32m2_t vssubu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m2_m))) +vuint32m2_t vssubu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m4))) +vuint32m4_t vssubu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m4_m))) +vuint32m4_t vssubu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m8))) +vuint32m8_t vssubu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m8_m))) +vuint32m8_t vssubu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32mf2))) +vuint32mf2_t vssubu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32mf2_m))) +vuint32mf2_t vssubu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m1))) +vuint64m1_t vssubu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m1_m))) +vuint64m1_t vssubu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m2))) +vuint64m2_t vssubu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m2_m))) +vuint64m2_t vssubu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m4))) +vuint64m4_t vssubu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m4_m))) +vuint64m4_t vssubu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m8))) +vuint64m8_t vssubu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m8_m))) +vuint64m8_t vssubu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m1))) +vuint8m1_t vssubu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m1_m))) +vuint8m1_t vssubu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m2))) +vuint8m2_t vssubu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m2_m))) +vuint8m2_t vssubu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m4))) +vuint8m4_t vssubu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m4_m))) +vuint8m4_t vssubu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m8))) +vuint8m8_t vssubu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m8_m))) +vuint8m8_t vssubu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8mf2))) +vuint8mf2_t vssubu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8mf2_m))) +vuint8mf2_t vssubu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8mf4))) +vuint8mf4_t vssubu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8mf4_m))) +vuint8mf4_t vssubu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8mf8))) +vuint8mf8_t vssubu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8mf8_m))) +vuint8mf8_t vssubu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m1))) +vuint16m1_t vssubu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m1_m))) +vuint16m1_t vssubu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m2))) +vuint16m2_t vssubu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m2_m))) +vuint16m2_t vssubu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m4))) +vuint16m4_t vssubu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m4_m))) +vuint16m4_t vssubu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m8))) +vuint16m8_t vssubu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m8_m))) +vuint16m8_t vssubu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16mf2))) +vuint16mf2_t vssubu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16mf2_m))) +vuint16mf2_t vssubu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16mf4))) +vuint16mf4_t vssubu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16mf4_m))) +vuint16mf4_t vssubu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m1))) +vuint32m1_t vssubu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m1_m))) +vuint32m1_t vssubu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m2))) +vuint32m2_t vssubu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m2_m))) +vuint32m2_t vssubu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m4))) +vuint32m4_t vssubu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m4_m))) +vuint32m4_t vssubu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m8))) +vuint32m8_t vssubu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m8_m))) +vuint32m8_t vssubu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32mf2))) +vuint32mf2_t vssubu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32mf2_m))) +vuint32mf2_t vssubu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m1))) +vuint64m1_t vssubu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m1_m))) +vuint64m1_t vssubu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m2))) +vuint64m2_t vssubu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m2_m))) +vuint64m2_t vssubu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m4))) +vuint64m4_t vssubu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m4_m))) +vuint64m4_t vssubu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m8))) +vuint64m8_t vssubu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m8_m))) +vuint64m8_t vssubu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m1))) +vint8m1_t vssub(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m1_m))) +vint8m1_t vssub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m2))) +vint8m2_t vssub(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m2_m))) +vint8m2_t vssub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m4))) +vint8m4_t vssub(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m4_m))) +vint8m4_t vssub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m8))) +vint8m8_t vssub(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m8_m))) +vint8m8_t vssub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8mf2))) +vint8mf2_t vssub(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8mf2_m))) +vint8mf2_t vssub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8mf4))) +vint8mf4_t vssub(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8mf4_m))) +vint8mf4_t vssub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8mf8))) +vint8mf8_t vssub(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8mf8_m))) +vint8mf8_t vssub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m1))) +vint16m1_t vssub(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m1_m))) +vint16m1_t vssub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m2))) +vint16m2_t vssub(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m2_m))) +vint16m2_t vssub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m4))) +vint16m4_t vssub(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m4_m))) +vint16m4_t vssub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m8))) +vint16m8_t vssub(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m8_m))) +vint16m8_t vssub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16mf2))) +vint16mf2_t vssub(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16mf2_m))) +vint16mf2_t vssub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16mf4))) +vint16mf4_t vssub(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16mf4_m))) +vint16mf4_t vssub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m1))) +vint32m1_t vssub(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m1_m))) +vint32m1_t vssub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m2))) +vint32m2_t vssub(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m2_m))) +vint32m2_t vssub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m4))) +vint32m4_t vssub(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m4_m))) +vint32m4_t vssub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m8))) +vint32m8_t vssub(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m8_m))) +vint32m8_t vssub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32mf2))) +vint32mf2_t vssub(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32mf2_m))) +vint32mf2_t vssub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m1))) +vint64m1_t vssub(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m1_m))) +vint64m1_t vssub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m2))) +vint64m2_t vssub(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m2_m))) +vint64m2_t vssub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m4))) +vint64m4_t vssub(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m4_m))) +vint64m4_t vssub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m8))) +vint64m8_t vssub(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m8_m))) +vint64m8_t vssub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m1))) +vint8m1_t vssub(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m1_m))) +vint8m1_t vssub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m2))) +vint8m2_t vssub(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m2_m))) +vint8m2_t vssub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m4))) +vint8m4_t vssub(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m4_m))) +vint8m4_t vssub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m8))) +vint8m8_t vssub(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m8_m))) +vint8m8_t vssub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8mf2))) +vint8mf2_t vssub(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8mf2_m))) +vint8mf2_t vssub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8mf4))) +vint8mf4_t vssub(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8mf4_m))) +vint8mf4_t vssub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8mf8))) +vint8mf8_t vssub(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8mf8_m))) +vint8mf8_t vssub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m1))) +vint16m1_t vssub(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m1_m))) +vint16m1_t vssub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m2))) +vint16m2_t vssub(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m2_m))) +vint16m2_t vssub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m4))) +vint16m4_t vssub(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m4_m))) +vint16m4_t vssub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m8))) +vint16m8_t vssub(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m8_m))) +vint16m8_t vssub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16mf2))) +vint16mf2_t vssub(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16mf2_m))) +vint16mf2_t vssub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16mf4))) +vint16mf4_t vssub(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16mf4_m))) +vint16mf4_t vssub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m1))) +vint32m1_t vssub(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m1_m))) +vint32m1_t vssub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m2))) +vint32m2_t vssub(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m2_m))) +vint32m2_t vssub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m4))) +vint32m4_t vssub(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m4_m))) +vint32m4_t vssub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m8))) +vint32m8_t vssub(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m8_m))) +vint32m8_t vssub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32mf2))) +vint32mf2_t vssub(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32mf2_m))) +vint32mf2_t vssub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m1))) +vint64m1_t vssub(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m1_m))) +vint64m1_t vssub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m2))) +vint64m2_t vssub(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m2_m))) +vint64m2_t vssub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m4))) +vint64m4_t vssub(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m4_m))) +vint64m4_t vssub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m8))) +vint64m8_t vssub(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m8_m))) +vint64m8_t vssub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m1))) +vuint8m1_t vaaddu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m1_m))) +vuint8m1_t vaaddu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m2))) +vuint8m2_t vaaddu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m2_m))) +vuint8m2_t vaaddu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m4))) +vuint8m4_t vaaddu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m4_m))) +vuint8m4_t vaaddu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m8))) +vuint8m8_t vaaddu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m8_m))) +vuint8m8_t vaaddu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8mf2))) +vuint8mf2_t vaaddu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8mf2_m))) +vuint8mf2_t vaaddu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8mf4))) +vuint8mf4_t vaaddu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8mf4_m))) +vuint8mf4_t vaaddu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8mf8))) +vuint8mf8_t vaaddu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8mf8_m))) +vuint8mf8_t vaaddu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m1))) +vuint16m1_t vaaddu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m1_m))) +vuint16m1_t vaaddu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m2))) +vuint16m2_t vaaddu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m2_m))) +vuint16m2_t vaaddu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m4))) +vuint16m4_t vaaddu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m4_m))) +vuint16m4_t vaaddu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m8))) +vuint16m8_t vaaddu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m8_m))) +vuint16m8_t vaaddu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16mf2))) +vuint16mf2_t vaaddu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16mf2_m))) +vuint16mf2_t vaaddu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16mf4))) +vuint16mf4_t vaaddu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16mf4_m))) +vuint16mf4_t vaaddu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m1))) +vuint32m1_t vaaddu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m1_m))) +vuint32m1_t vaaddu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m2))) +vuint32m2_t vaaddu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m2_m))) +vuint32m2_t vaaddu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m4))) +vuint32m4_t vaaddu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m4_m))) +vuint32m4_t vaaddu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m8))) +vuint32m8_t vaaddu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m8_m))) +vuint32m8_t vaaddu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32mf2))) +vuint32mf2_t vaaddu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32mf2_m))) +vuint32mf2_t vaaddu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m1))) +vuint64m1_t vaaddu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m1_m))) +vuint64m1_t vaaddu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m2))) +vuint64m2_t vaaddu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m2_m))) +vuint64m2_t vaaddu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m4))) +vuint64m4_t vaaddu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m4_m))) +vuint64m4_t vaaddu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m8))) +vuint64m8_t vaaddu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m8_m))) +vuint64m8_t vaaddu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m1))) +vuint8m1_t vaaddu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m1_m))) +vuint8m1_t vaaddu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m2))) +vuint8m2_t vaaddu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m2_m))) +vuint8m2_t vaaddu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m4))) +vuint8m4_t vaaddu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m4_m))) +vuint8m4_t vaaddu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m8))) +vuint8m8_t vaaddu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m8_m))) +vuint8m8_t vaaddu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8mf2))) +vuint8mf2_t vaaddu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8mf2_m))) +vuint8mf2_t vaaddu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8mf4))) +vuint8mf4_t vaaddu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8mf4_m))) +vuint8mf4_t vaaddu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8mf8))) +vuint8mf8_t vaaddu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8mf8_m))) +vuint8mf8_t vaaddu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m1))) +vuint16m1_t vaaddu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m1_m))) +vuint16m1_t vaaddu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m2))) +vuint16m2_t vaaddu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m2_m))) +vuint16m2_t vaaddu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m4))) +vuint16m4_t vaaddu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m4_m))) +vuint16m4_t vaaddu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m8))) +vuint16m8_t vaaddu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m8_m))) +vuint16m8_t vaaddu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16mf2))) +vuint16mf2_t vaaddu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16mf2_m))) +vuint16mf2_t vaaddu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16mf4))) +vuint16mf4_t vaaddu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16mf4_m))) +vuint16mf4_t vaaddu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m1))) +vuint32m1_t vaaddu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m1_m))) +vuint32m1_t vaaddu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m2))) +vuint32m2_t vaaddu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m2_m))) +vuint32m2_t vaaddu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m4))) +vuint32m4_t vaaddu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m4_m))) +vuint32m4_t vaaddu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m8))) +vuint32m8_t vaaddu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m8_m))) +vuint32m8_t vaaddu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32mf2))) +vuint32mf2_t vaaddu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32mf2_m))) +vuint32mf2_t vaaddu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m1))) +vuint64m1_t vaaddu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m1_m))) +vuint64m1_t vaaddu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m2))) +vuint64m2_t vaaddu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m2_m))) +vuint64m2_t vaaddu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m4))) +vuint64m4_t vaaddu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m4_m))) +vuint64m4_t vaaddu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m8))) +vuint64m8_t vaaddu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m8_m))) +vuint64m8_t vaaddu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m1))) +vint8m1_t vaadd(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m1_m))) +vint8m1_t vaadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m2))) +vint8m2_t vaadd(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m2_m))) +vint8m2_t vaadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m4))) +vint8m4_t vaadd(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m4_m))) +vint8m4_t vaadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m8))) +vint8m8_t vaadd(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m8_m))) +vint8m8_t vaadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8mf2))) +vint8mf2_t vaadd(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8mf2_m))) +vint8mf2_t vaadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8mf4))) +vint8mf4_t vaadd(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8mf4_m))) +vint8mf4_t vaadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8mf8))) +vint8mf8_t vaadd(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8mf8_m))) +vint8mf8_t vaadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m1))) +vint16m1_t vaadd(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m1_m))) +vint16m1_t vaadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m2))) +vint16m2_t vaadd(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m2_m))) +vint16m2_t vaadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m4))) +vint16m4_t vaadd(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m4_m))) +vint16m4_t vaadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m8))) +vint16m8_t vaadd(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m8_m))) +vint16m8_t vaadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16mf2))) +vint16mf2_t vaadd(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16mf2_m))) +vint16mf2_t vaadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16mf4))) +vint16mf4_t vaadd(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16mf4_m))) +vint16mf4_t vaadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m1))) +vint32m1_t vaadd(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m1_m))) +vint32m1_t vaadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m2))) +vint32m2_t vaadd(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m2_m))) +vint32m2_t vaadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m4))) +vint32m4_t vaadd(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m4_m))) +vint32m4_t vaadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m8))) +vint32m8_t vaadd(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m8_m))) +vint32m8_t vaadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32mf2))) +vint32mf2_t vaadd(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32mf2_m))) +vint32mf2_t vaadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m1))) +vint64m1_t vaadd(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m1_m))) +vint64m1_t vaadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m2))) +vint64m2_t vaadd(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m2_m))) +vint64m2_t vaadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m4))) +vint64m4_t vaadd(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m4_m))) +vint64m4_t vaadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m8))) +vint64m8_t vaadd(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m8_m))) +vint64m8_t vaadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m1))) +vint8m1_t vaadd(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m1_m))) +vint8m1_t vaadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m2))) +vint8m2_t vaadd(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m2_m))) +vint8m2_t vaadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m4))) +vint8m4_t vaadd(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m4_m))) +vint8m4_t vaadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m8))) +vint8m8_t vaadd(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m8_m))) +vint8m8_t vaadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8mf2))) +vint8mf2_t vaadd(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8mf2_m))) +vint8mf2_t vaadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8mf4))) +vint8mf4_t vaadd(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8mf4_m))) +vint8mf4_t vaadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8mf8))) +vint8mf8_t vaadd(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8mf8_m))) +vint8mf8_t vaadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m1))) +vint16m1_t vaadd(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m1_m))) +vint16m1_t vaadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m2))) +vint16m2_t vaadd(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m2_m))) +vint16m2_t vaadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m4))) +vint16m4_t vaadd(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m4_m))) +vint16m4_t vaadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m8))) +vint16m8_t vaadd(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m8_m))) +vint16m8_t vaadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16mf2))) +vint16mf2_t vaadd(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16mf2_m))) +vint16mf2_t vaadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16mf4))) +vint16mf4_t vaadd(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16mf4_m))) +vint16mf4_t vaadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m1))) +vint32m1_t vaadd(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m1_m))) +vint32m1_t vaadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m2))) +vint32m2_t vaadd(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m2_m))) +vint32m2_t vaadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m4))) +vint32m4_t vaadd(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m4_m))) +vint32m4_t vaadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m8))) +vint32m8_t vaadd(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m8_m))) +vint32m8_t vaadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32mf2))) +vint32mf2_t vaadd(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32mf2_m))) +vint32mf2_t vaadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m1))) +vint64m1_t vaadd(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m1_m))) +vint64m1_t vaadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m2))) +vint64m2_t vaadd(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m2_m))) +vint64m2_t vaadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m4))) +vint64m4_t vaadd(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m4_m))) +vint64m4_t vaadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m8))) +vint64m8_t vaadd(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m8_m))) +vint64m8_t vaadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m1))) +vuint8m1_t vasubu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m1_m))) +vuint8m1_t vasubu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m2))) +vuint8m2_t vasubu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m2_m))) +vuint8m2_t vasubu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m4))) +vuint8m4_t vasubu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m4_m))) +vuint8m4_t vasubu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m8))) +vuint8m8_t vasubu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m8_m))) +vuint8m8_t vasubu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8mf2))) +vuint8mf2_t vasubu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8mf2_m))) +vuint8mf2_t vasubu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8mf4))) +vuint8mf4_t vasubu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8mf4_m))) +vuint8mf4_t vasubu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8mf8))) +vuint8mf8_t vasubu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8mf8_m))) +vuint8mf8_t vasubu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m1))) +vuint16m1_t vasubu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m1_m))) +vuint16m1_t vasubu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m2))) +vuint16m2_t vasubu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m2_m))) +vuint16m2_t vasubu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m4))) +vuint16m4_t vasubu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m4_m))) +vuint16m4_t vasubu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m8))) +vuint16m8_t vasubu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m8_m))) +vuint16m8_t vasubu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16mf2))) +vuint16mf2_t vasubu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16mf2_m))) +vuint16mf2_t vasubu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16mf4))) +vuint16mf4_t vasubu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16mf4_m))) +vuint16mf4_t vasubu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m1))) +vuint32m1_t vasubu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m1_m))) +vuint32m1_t vasubu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m2))) +vuint32m2_t vasubu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m2_m))) +vuint32m2_t vasubu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m4))) +vuint32m4_t vasubu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m4_m))) +vuint32m4_t vasubu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m8))) +vuint32m8_t vasubu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m8_m))) +vuint32m8_t vasubu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32mf2))) +vuint32mf2_t vasubu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32mf2_m))) +vuint32mf2_t vasubu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m1))) +vuint64m1_t vasubu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m1_m))) +vuint64m1_t vasubu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m2))) +vuint64m2_t vasubu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m2_m))) +vuint64m2_t vasubu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m4))) +vuint64m4_t vasubu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m4_m))) +vuint64m4_t vasubu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m8))) +vuint64m8_t vasubu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m8_m))) +vuint64m8_t vasubu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m1))) +vuint8m1_t vasubu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m1_m))) +vuint8m1_t vasubu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m2))) +vuint8m2_t vasubu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m2_m))) +vuint8m2_t vasubu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m4))) +vuint8m4_t vasubu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m4_m))) +vuint8m4_t vasubu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m8))) +vuint8m8_t vasubu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m8_m))) +vuint8m8_t vasubu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8mf2))) +vuint8mf2_t vasubu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8mf2_m))) +vuint8mf2_t vasubu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8mf4))) +vuint8mf4_t vasubu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8mf4_m))) +vuint8mf4_t vasubu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8mf8))) +vuint8mf8_t vasubu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8mf8_m))) +vuint8mf8_t vasubu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m1))) +vuint16m1_t vasubu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m1_m))) +vuint16m1_t vasubu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m2))) +vuint16m2_t vasubu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m2_m))) +vuint16m2_t vasubu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m4))) +vuint16m4_t vasubu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m4_m))) +vuint16m4_t vasubu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m8))) +vuint16m8_t vasubu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m8_m))) +vuint16m8_t vasubu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16mf2))) +vuint16mf2_t vasubu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16mf2_m))) +vuint16mf2_t vasubu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16mf4))) +vuint16mf4_t vasubu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16mf4_m))) +vuint16mf4_t vasubu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m1))) +vuint32m1_t vasubu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m1_m))) +vuint32m1_t vasubu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m2))) +vuint32m2_t vasubu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m2_m))) +vuint32m2_t vasubu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m4))) +vuint32m4_t vasubu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m4_m))) +vuint32m4_t vasubu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m8))) +vuint32m8_t vasubu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m8_m))) +vuint32m8_t vasubu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32mf2))) +vuint32mf2_t vasubu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32mf2_m))) +vuint32mf2_t vasubu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m1))) +vuint64m1_t vasubu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m1_m))) +vuint64m1_t vasubu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m2))) +vuint64m2_t vasubu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m2_m))) +vuint64m2_t vasubu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m4))) +vuint64m4_t vasubu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m4_m))) +vuint64m4_t vasubu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m8))) +vuint64m8_t vasubu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m8_m))) +vuint64m8_t vasubu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m1))) +void vsuxei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m1_m))) +void vsuxei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m2))) +void vsuxei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m2_m))) +void vsuxei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m4))) +void vsuxei8(int8_t * op0, vuint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m4_m))) +void vsuxei8(vbool2_t op0, int8_t * op1, vuint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m8))) +void vsuxei8(int8_t * op0, vuint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m8_m))) +void vsuxei8(vbool1_t op0, int8_t * op1, vuint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8mf2))) +void vsuxei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8mf2_m))) +void vsuxei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8mf4))) +void vsuxei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8mf4_m))) +void vsuxei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8mf8))) +void vsuxei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8mf8_m))) +void vsuxei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32m1))) +void vsuxei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32m1_m))) +void vsuxei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32m2))) +void vsuxei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32m2_m))) +void vsuxei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32m4))) +void vsuxei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32m4_m))) +void vsuxei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32mf2))) +void vsuxei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32mf2_m))) +void vsuxei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m1))) +vint8m1_t vasub(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m1_m))) +vint8m1_t vasub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m2))) +vint8m2_t vasub(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m2_m))) +vint8m2_t vasub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m4))) +vint8m4_t vasub(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m4_m))) +vint8m4_t vasub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m8))) +vint8m8_t vasub(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m8_m))) +vint8m8_t vasub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8mf2))) +vint8mf2_t vasub(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8mf2_m))) +vint8mf2_t vasub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8mf4))) +vint8mf4_t vasub(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8mf4_m))) +vint8mf4_t vasub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8mf8))) +vint8mf8_t vasub(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8mf8_m))) +vint8mf8_t vasub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m1))) +vint16m1_t vasub(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m1_m))) +vint16m1_t vasub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m2))) +vint16m2_t vasub(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m2_m))) +vint16m2_t vasub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m4))) +vint16m4_t vasub(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m4_m))) +vint16m4_t vasub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m8))) +vint16m8_t vasub(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m8_m))) +vint16m8_t vasub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16mf2))) +vint16mf2_t vasub(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16mf2_m))) +vint16mf2_t vasub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16mf4))) +vint16mf4_t vasub(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16mf4_m))) +vint16mf4_t vasub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m1))) +vint32m1_t vasub(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m1_m))) +vint32m1_t vasub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m2))) +vint32m2_t vasub(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m2_m))) +vint32m2_t vasub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m4))) +vint32m4_t vasub(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m4_m))) +vint32m4_t vasub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m8))) +vint32m8_t vasub(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m8_m))) +vint32m8_t vasub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32mf2))) +vint32mf2_t vasub(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32mf2_m))) +vint32mf2_t vasub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m1))) +vint64m1_t vasub(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m1_m))) +vint64m1_t vasub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m2))) +vint64m2_t vasub(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m2_m))) +vint64m2_t vasub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m4))) +vint64m4_t vasub(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m4_m))) +vint64m4_t vasub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m8))) +vint64m8_t vasub(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m8_m))) +vint64m8_t vasub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m1))) +vint8m1_t vasub(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m1_m))) +vint8m1_t vasub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m2))) +vint8m2_t vasub(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m2_m))) +vint8m2_t vasub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m4))) +vint8m4_t vasub(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m4_m))) +vint8m4_t vasub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m8))) +vint8m8_t vasub(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m8_m))) +vint8m8_t vasub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8mf2))) +vint8mf2_t vasub(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8mf2_m))) +vint8mf2_t vasub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8mf4))) +vint8mf4_t vasub(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8mf4_m))) +vint8mf4_t vasub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8mf8))) +vint8mf8_t vasub(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8mf8_m))) +vint8mf8_t vasub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m1))) +vint16m1_t vasub(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m1_m))) +vint16m1_t vasub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m2))) +vint16m2_t vasub(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m2_m))) +vint16m2_t vasub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m4))) +vint16m4_t vasub(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m4_m))) +vint16m4_t vasub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m8))) +vint16m8_t vasub(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m8_m))) +vint16m8_t vasub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16mf2))) +vint16mf2_t vasub(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16mf2_m))) +vint16mf2_t vasub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16mf4))) +vint16mf4_t vasub(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16mf4_m))) +vint16mf4_t vasub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m1))) +vint32m1_t vasub(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m1_m))) +vint32m1_t vasub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m2))) +vint32m2_t vasub(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m2_m))) +vint32m2_t vasub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m4))) +vint32m4_t vasub(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m4_m))) +vint32m4_t vasub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m8))) +vint32m8_t vasub(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m8_m))) +vint32m8_t vasub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32mf2))) +vint32mf2_t vasub(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32mf2_m))) +vint32mf2_t vasub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m1))) +vint64m1_t vasub(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m1_m))) +vint64m1_t vasub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m2))) +vint64m2_t vasub(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m2_m))) +vint64m2_t vasub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m4))) +vint64m4_t vasub(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m4_m))) +vint64m4_t vasub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m8))) +vint64m8_t vasub(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m8_m))) +vint64m8_t vasub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m1))) +vint8m1_t vsmul(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m1_m))) +vint8m1_t vsmul(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m2))) +vint8m2_t vsmul(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m2_m))) +vint8m2_t vsmul(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m4))) +vint8m4_t vsmul(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m4_m))) +vint8m4_t vsmul(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m8))) +vint8m8_t vsmul(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m8_m))) +vint8m8_t vsmul(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8mf2))) +vint8mf2_t vsmul(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8mf2_m))) +vint8mf2_t vsmul(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8mf4))) +vint8mf4_t vsmul(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8mf4_m))) +vint8mf4_t vsmul(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8mf8))) +vint8mf8_t vsmul(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8mf8_m))) +vint8mf8_t vsmul(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m1))) +vint16m1_t vsmul(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m1_m))) +vint16m1_t vsmul(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m2))) +vint16m2_t vsmul(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m2_m))) +vint16m2_t vsmul(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m4))) +vint16m4_t vsmul(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m4_m))) +vint16m4_t vsmul(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m8))) +vint16m8_t vsmul(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m8_m))) +vint16m8_t vsmul(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16mf2))) +vint16mf2_t vsmul(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16mf2_m))) +vint16mf2_t vsmul(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16mf4))) +vint16mf4_t vsmul(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16mf4_m))) +vint16mf4_t vsmul(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m1))) +vint32m1_t vsmul(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m1_m))) +vint32m1_t vsmul(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m2))) +vint32m2_t vsmul(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m2_m))) +vint32m2_t vsmul(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m4))) +vint32m4_t vsmul(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m4_m))) +vint32m4_t vsmul(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m8))) +vint32m8_t vsmul(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m8_m))) +vint32m8_t vsmul(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32mf2))) +vint32mf2_t vsmul(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32mf2_m))) +vint32mf2_t vsmul(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m1))) +vint64m1_t vsmul(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m1_m))) +vint64m1_t vsmul(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m2))) +vint64m2_t vsmul(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m2_m))) +vint64m2_t vsmul(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m4))) +vint64m4_t vsmul(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m4_m))) +vint64m4_t vsmul(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m8))) +vint64m8_t vsmul(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m8_m))) +vint64m8_t vsmul(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m1))) +vint8m1_t vsmul(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m1_m))) +vint8m1_t vsmul(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m2))) +vint8m2_t vsmul(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m2_m))) +vint8m2_t vsmul(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m4))) +vint8m4_t vsmul(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m4_m))) +vint8m4_t vsmul(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m8))) +vint8m8_t vsmul(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m8_m))) +vint8m8_t vsmul(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8mf2))) +vint8mf2_t vsmul(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8mf2_m))) +vint8mf2_t vsmul(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8mf4))) +vint8mf4_t vsmul(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8mf4_m))) +vint8mf4_t vsmul(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8mf8))) +vint8mf8_t vsmul(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8mf8_m))) +vint8mf8_t vsmul(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m1))) +vint16m1_t vsmul(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m1_m))) +vint16m1_t vsmul(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m2))) +vint16m2_t vsmul(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m2_m))) +vint16m2_t vsmul(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m4))) +vint16m4_t vsmul(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m4_m))) +vint16m4_t vsmul(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m8))) +vint16m8_t vsmul(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m8_m))) +vint16m8_t vsmul(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16mf2))) +vint16mf2_t vsmul(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16mf2_m))) +vint16mf2_t vsmul(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16mf4))) +vint16mf4_t vsmul(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16mf4_m))) +vint16mf4_t vsmul(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m1))) +vint32m1_t vsmul(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m1_m))) +vint32m1_t vsmul(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m2))) +vint32m2_t vsmul(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m2_m))) +vint32m2_t vsmul(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m4))) +vint32m4_t vsmul(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m4_m))) +vint32m4_t vsmul(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m8))) +vint32m8_t vsmul(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m8_m))) +vint32m8_t vsmul(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32mf2))) +vint32mf2_t vsmul(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32mf2_m))) +vint32mf2_t vsmul(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m1))) +vint64m1_t vsmul(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m1_m))) +vint64m1_t vsmul(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m2))) +vint64m2_t vsmul(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m2_m))) +vint64m2_t vsmul(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m4))) +vint64m4_t vsmul(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m4_m))) +vint64m4_t vsmul(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m8))) +vint64m8_t vsmul(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m8_m))) +vint64m8_t vsmul(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m1))) +vuint8m1_t vssrl(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m1_m))) +vuint8m1_t vssrl(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m2))) +vuint8m2_t vssrl(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m2_m))) +vuint8m2_t vssrl(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m4))) +vuint8m4_t vssrl(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m4_m))) +vuint8m4_t vssrl(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m8))) +vuint8m8_t vssrl(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m8_m))) +vuint8m8_t vssrl(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8mf2))) +vuint8mf2_t vssrl(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8mf2_m))) +vuint8mf2_t vssrl(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8mf4))) +vuint8mf4_t vssrl(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8mf4_m))) +vuint8mf4_t vssrl(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8mf8))) +vuint8mf8_t vssrl(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8mf8_m))) +vuint8mf8_t vssrl(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m1))) +vuint16m1_t vssrl(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m1_m))) +vuint16m1_t vssrl(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m2))) +vuint16m2_t vssrl(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m2_m))) +vuint16m2_t vssrl(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m4))) +vuint16m4_t vssrl(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m4_m))) +vuint16m4_t vssrl(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m8))) +vuint16m8_t vssrl(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m8_m))) +vuint16m8_t vssrl(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16mf2))) +vuint16mf2_t vssrl(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16mf2_m))) +vuint16mf2_t vssrl(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16mf4))) +vuint16mf4_t vssrl(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16mf4_m))) +vuint16mf4_t vssrl(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m1))) +vuint32m1_t vssrl(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m1_m))) +vuint32m1_t vssrl(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m2))) +vuint32m2_t vssrl(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m2_m))) +vuint32m2_t vssrl(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m4))) +vuint32m4_t vssrl(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m4_m))) +vuint32m4_t vssrl(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m8))) +vuint32m8_t vssrl(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m8_m))) +vuint32m8_t vssrl(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32mf2))) +vuint32mf2_t vssrl(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32mf2_m))) +vuint32mf2_t vssrl(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m1))) +vuint64m1_t vssrl(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m1_m))) +vuint64m1_t vssrl(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m2))) +vuint64m2_t vssrl(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m2_m))) +vuint64m2_t vssrl(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m4))) +vuint64m4_t vssrl(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m4_m))) +vuint64m4_t vssrl(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m8))) +vuint64m8_t vssrl(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m8_m))) +vuint64m8_t vssrl(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m1))) +vuint8m1_t vssrl(vuint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m1_m))) +vuint8m1_t vssrl(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m2))) +vuint8m2_t vssrl(vuint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m2_m))) +vuint8m2_t vssrl(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m4))) +vuint8m4_t vssrl(vuint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m4_m))) +vuint8m4_t vssrl(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m8))) +vuint8m8_t vssrl(vuint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m8_m))) +vuint8m8_t vssrl(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8mf2))) +vuint8mf2_t vssrl(vuint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8mf2_m))) +vuint8mf2_t vssrl(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8mf4))) +vuint8mf4_t vssrl(vuint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8mf4_m))) +vuint8mf4_t vssrl(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8mf8))) +vuint8mf8_t vssrl(vuint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8mf8_m))) +vuint8mf8_t vssrl(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m1))) +vuint16m1_t vssrl(vuint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m1_m))) +vuint16m1_t vssrl(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m2))) +vuint16m2_t vssrl(vuint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m2_m))) +vuint16m2_t vssrl(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m4))) +vuint16m4_t vssrl(vuint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m4_m))) +vuint16m4_t vssrl(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m8))) +vuint16m8_t vssrl(vuint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m8_m))) +vuint16m8_t vssrl(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16mf2))) +vuint16mf2_t vssrl(vuint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16mf2_m))) +vuint16mf2_t vssrl(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16mf4))) +vuint16mf4_t vssrl(vuint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16mf4_m))) +vuint16mf4_t vssrl(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m1))) +vuint32m1_t vssrl(vuint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m1_m))) +vuint32m1_t vssrl(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m2))) +vuint32m2_t vssrl(vuint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m2_m))) +vuint32m2_t vssrl(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m4))) +vuint32m4_t vssrl(vuint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m4_m))) +vuint32m4_t vssrl(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m8))) +vuint32m8_t vssrl(vuint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m8_m))) +vuint32m8_t vssrl(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32mf2))) +vuint32mf2_t vssrl(vuint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32mf2_m))) +vuint32mf2_t vssrl(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m1))) +vuint64m1_t vssrl(vuint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m1_m))) +vuint64m1_t vssrl(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m2))) +vuint64m2_t vssrl(vuint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m2_m))) +vuint64m2_t vssrl(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m4))) +vuint64m4_t vssrl(vuint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m4_m))) +vuint64m4_t vssrl(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m8))) +vuint64m8_t vssrl(vuint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m8_m))) +vuint64m8_t vssrl(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m1))) +vint8m1_t vssra(vint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m1_m))) +vint8m1_t vssra(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m2))) +vint8m2_t vssra(vint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m2_m))) +vint8m2_t vssra(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m4))) +vint8m4_t vssra(vint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m4_m))) +vint8m4_t vssra(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m8))) +vint8m8_t vssra(vint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m8_m))) +vint8m8_t vssra(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8mf2))) +vint8mf2_t vssra(vint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8mf2_m))) +vint8mf2_t vssra(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8mf4))) +vint8mf4_t vssra(vint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8mf4_m))) +vint8mf4_t vssra(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8mf8))) +vint8mf8_t vssra(vint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8mf8_m))) +vint8mf8_t vssra(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m1))) +vint16m1_t vssra(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m1_m))) +vint16m1_t vssra(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m2))) +vint16m2_t vssra(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m2_m))) +vint16m2_t vssra(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m4))) +vint16m4_t vssra(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m4_m))) +vint16m4_t vssra(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m8))) +vint16m8_t vssra(vint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m8_m))) +vint16m8_t vssra(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16mf2))) +vint16mf2_t vssra(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16mf2_m))) +vint16mf2_t vssra(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16mf4))) +vint16mf4_t vssra(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16mf4_m))) +vint16mf4_t vssra(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m1))) +vint32m1_t vssra(vint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m1_m))) +vint32m1_t vssra(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m2))) +vint32m2_t vssra(vint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m2_m))) +vint32m2_t vssra(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m4))) +vint32m4_t vssra(vint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m4_m))) +vint32m4_t vssra(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m8))) +vint32m8_t vssra(vint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m8_m))) +vint32m8_t vssra(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32mf2))) +vint32mf2_t vssra(vint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32mf2_m))) +vint32mf2_t vssra(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m1))) +vint64m1_t vssra(vint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m1_m))) +vint64m1_t vssra(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m2))) +vint64m2_t vssra(vint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m2_m))) +vint64m2_t vssra(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m4))) +vint64m4_t vssra(vint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m4_m))) +vint64m4_t vssra(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m8))) +vint64m8_t vssra(vint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m8_m))) +vint64m8_t vssra(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m1))) +vint8m1_t vssra(vint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m1_m))) +vint8m1_t vssra(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m2))) +vint8m2_t vssra(vint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m2_m))) +vint8m2_t vssra(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m4))) +vint8m4_t vssra(vint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m4_m))) +vint8m4_t vssra(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m8))) +vint8m8_t vssra(vint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m8_m))) +vint8m8_t vssra(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8mf2))) +vint8mf2_t vssra(vint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8mf2_m))) +vint8mf2_t vssra(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8mf4))) +vint8mf4_t vssra(vint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8mf4_m))) +vint8mf4_t vssra(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8mf8))) +vint8mf8_t vssra(vint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8mf8_m))) +vint8mf8_t vssra(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m1))) +vint16m1_t vssra(vint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m1_m))) +vint16m1_t vssra(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m2))) +vint16m2_t vssra(vint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m2_m))) +vint16m2_t vssra(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m4))) +vint16m4_t vssra(vint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m4_m))) +vint16m4_t vssra(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m8))) +vint16m8_t vssra(vint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m8_m))) +vint16m8_t vssra(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16mf2))) +vint16mf2_t vssra(vint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16mf2_m))) +vint16mf2_t vssra(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16mf4))) +vint16mf4_t vssra(vint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16mf4_m))) +vint16mf4_t vssra(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m1))) +vint32m1_t vssra(vint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m1_m))) +vint32m1_t vssra(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m2))) +vint32m2_t vssra(vint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m2_m))) +vint32m2_t vssra(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m4))) +vint32m4_t vssra(vint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m4_m))) +vint32m4_t vssra(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m8))) +vint32m8_t vssra(vint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m8_m))) +vint32m8_t vssra(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32mf2))) +vint32mf2_t vssra(vint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32mf2_m))) +vint32mf2_t vssra(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m1))) +vint64m1_t vssra(vint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m1_m))) +vint64m1_t vssra(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m2))) +vint64m2_t vssra(vint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m2_m))) +vint64m2_t vssra(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m4))) +vint64m4_t vssra(vint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m4_m))) +vint64m4_t vssra(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m8))) +vint64m8_t vssra(vint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m8_m))) +vint64m8_t vssra(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8m1))) +vuint8m1_t vnclipu(vuint16m2_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8m1_m))) +vuint8m1_t vnclipu(vbool8_t op0, vuint8m1_t op1, vuint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8m2))) +vuint8m2_t vnclipu(vuint16m4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8m2_m))) +vuint8m2_t vnclipu(vbool4_t op0, vuint8m2_t op1, vuint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8m4))) +vuint8m4_t vnclipu(vuint16m8_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8m4_m))) +vuint8m4_t vnclipu(vbool2_t op0, vuint8m4_t op1, vuint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8mf2))) +vuint8mf2_t vnclipu(vuint16m1_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8mf2_m))) +vuint8mf2_t vnclipu(vbool16_t op0, vuint8mf2_t op1, vuint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8mf4))) +vuint8mf4_t vnclipu(vuint16mf2_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8mf4_m))) +vuint8mf4_t vnclipu(vbool32_t op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8mf8))) +vuint8mf8_t vnclipu(vuint16mf4_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8mf8_m))) +vuint8mf8_t vnclipu(vbool64_t op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16m1))) +vuint16m1_t vnclipu(vuint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16m1_m))) +vuint16m1_t vnclipu(vbool16_t op0, vuint16m1_t op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16m2))) +vuint16m2_t vnclipu(vuint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16m2_m))) +vuint16m2_t vnclipu(vbool8_t op0, vuint16m2_t op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16m4))) +vuint16m4_t vnclipu(vuint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16m4_m))) +vuint16m4_t vnclipu(vbool4_t op0, vuint16m4_t op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16mf2))) +vuint16mf2_t vnclipu(vuint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16mf2_m))) +vuint16mf2_t vnclipu(vbool32_t op0, vuint16mf2_t op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16mf4))) +vuint16mf4_t vnclipu(vuint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16mf4_m))) +vuint16mf4_t vnclipu(vbool64_t op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32m1))) +vuint32m1_t vnclipu(vuint64m2_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32m1_m))) +vuint32m1_t vnclipu(vbool32_t op0, vuint32m1_t op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32m2))) +vuint32m2_t vnclipu(vuint64m4_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32m2_m))) +vuint32m2_t vnclipu(vbool16_t op0, vuint32m2_t op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32m4))) +vuint32m4_t vnclipu(vuint64m8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32m4_m))) +vuint32m4_t vnclipu(vbool8_t op0, vuint32m4_t op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32mf2))) +vuint32mf2_t vnclipu(vuint64m1_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32mf2_m))) +vuint32mf2_t vnclipu(vbool64_t op0, vuint32mf2_t op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8m1))) +vuint8m1_t vnclipu(vuint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8m1_m))) +vuint8m1_t vnclipu(vbool8_t op0, vuint8m1_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8m2))) +vuint8m2_t vnclipu(vuint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8m2_m))) +vuint8m2_t vnclipu(vbool4_t op0, vuint8m2_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8m4))) +vuint8m4_t vnclipu(vuint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8m4_m))) +vuint8m4_t vnclipu(vbool2_t op0, vuint8m4_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8mf2))) +vuint8mf2_t vnclipu(vuint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8mf2_m))) +vuint8mf2_t vnclipu(vbool16_t op0, vuint8mf2_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8mf4))) +vuint8mf4_t vnclipu(vuint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8mf4_m))) +vuint8mf4_t vnclipu(vbool32_t op0, vuint8mf4_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8mf8))) +vuint8mf8_t vnclipu(vuint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8mf8_m))) +vuint8mf8_t vnclipu(vbool64_t op0, vuint8mf8_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16m1))) +vuint16m1_t vnclipu(vuint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16m1_m))) +vuint16m1_t vnclipu(vbool16_t op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16m2))) +vuint16m2_t vnclipu(vuint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16m2_m))) +vuint16m2_t vnclipu(vbool8_t op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16m4))) +vuint16m4_t vnclipu(vuint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16m4_m))) +vuint16m4_t vnclipu(vbool4_t op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16mf2))) +vuint16mf2_t vnclipu(vuint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16mf2_m))) +vuint16mf2_t vnclipu(vbool32_t op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16mf4))) +vuint16mf4_t vnclipu(vuint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16mf4_m))) +vuint16mf4_t vnclipu(vbool64_t op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32m1))) +vuint32m1_t vnclipu(vuint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32m1_m))) +vuint32m1_t vnclipu(vbool32_t op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32m2))) +vuint32m2_t vnclipu(vuint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32m2_m))) +vuint32m2_t vnclipu(vbool16_t op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32m4))) +vuint32m4_t vnclipu(vuint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32m4_m))) +vuint32m4_t vnclipu(vbool8_t op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32mf2))) +vuint32mf2_t vnclipu(vuint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32mf2_m))) +vuint32mf2_t vnclipu(vbool64_t op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32m1))) +void vsuxei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32m1_m))) +void vsuxei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32m2))) +void vsuxei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32m2_m))) +void vsuxei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32m4))) +void vsuxei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32m4_m))) +void vsuxei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32mf2))) +void vsuxei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32mf2_m))) +void vsuxei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8m1))) +vint8m1_t vnclip(vint16m2_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8m1_m))) +vint8m1_t vnclip(vbool8_t op0, vint8m1_t op1, vint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8m2))) +vint8m2_t vnclip(vint16m4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8m2_m))) +vint8m2_t vnclip(vbool4_t op0, vint8m2_t op1, vint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8m4))) +vint8m4_t vnclip(vint16m8_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8m4_m))) +vint8m4_t vnclip(vbool2_t op0, vint8m4_t op1, vint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8mf2))) +vint8mf2_t vnclip(vint16m1_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8mf2_m))) +vint8mf2_t vnclip(vbool16_t op0, vint8mf2_t op1, vint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8mf4))) +vint8mf4_t vnclip(vint16mf2_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8mf4_m))) +vint8mf4_t vnclip(vbool32_t op0, vint8mf4_t op1, vint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8mf8))) +vint8mf8_t vnclip(vint16mf4_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8mf8_m))) +vint8mf8_t vnclip(vbool64_t op0, vint8mf8_t op1, vint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16m1))) +vint16m1_t vnclip(vint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16m1_m))) +vint16m1_t vnclip(vbool16_t op0, vint16m1_t op1, vint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16m2))) +vint16m2_t vnclip(vint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16m2_m))) +vint16m2_t vnclip(vbool8_t op0, vint16m2_t op1, vint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16m4))) +vint16m4_t vnclip(vint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16m4_m))) +vint16m4_t vnclip(vbool4_t op0, vint16m4_t op1, vint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16mf2))) +vint16mf2_t vnclip(vint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16mf2_m))) +vint16mf2_t vnclip(vbool32_t op0, vint16mf2_t op1, vint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16mf4))) +vint16mf4_t vnclip(vint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16mf4_m))) +vint16mf4_t vnclip(vbool64_t op0, vint16mf4_t op1, vint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32m1))) +vint32m1_t vnclip(vint64m2_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32m1_m))) +vint32m1_t vnclip(vbool32_t op0, vint32m1_t op1, vint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32m2))) +vint32m2_t vnclip(vint64m4_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32m2_m))) +vint32m2_t vnclip(vbool16_t op0, vint32m2_t op1, vint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32m4))) +vint32m4_t vnclip(vint64m8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32m4_m))) +vint32m4_t vnclip(vbool8_t op0, vint32m4_t op1, vint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32mf2))) +vint32mf2_t vnclip(vint64m1_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32mf2_m))) +vint32mf2_t vnclip(vbool64_t op0, vint32mf2_t op1, vint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8m1))) +vint8m1_t vnclip(vint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8m1_m))) +vint8m1_t vnclip(vbool8_t op0, vint8m1_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8m2))) +vint8m2_t vnclip(vint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8m2_m))) +vint8m2_t vnclip(vbool4_t op0, vint8m2_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8m4))) +vint8m4_t vnclip(vint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8m4_m))) +vint8m4_t vnclip(vbool2_t op0, vint8m4_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8mf2))) +vint8mf2_t vnclip(vint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8mf2_m))) +vint8mf2_t vnclip(vbool16_t op0, vint8mf2_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8mf4))) +vint8mf4_t vnclip(vint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8mf4_m))) +vint8mf4_t vnclip(vbool32_t op0, vint8mf4_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8mf8))) +vint8mf8_t vnclip(vint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8mf8_m))) +vint8mf8_t vnclip(vbool64_t op0, vint8mf8_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16m1))) +vint16m1_t vnclip(vint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16m1_m))) +vint16m1_t vnclip(vbool16_t op0, vint16m1_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16m2))) +vint16m2_t vnclip(vint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16m2_m))) +vint16m2_t vnclip(vbool8_t op0, vint16m2_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16m4))) +vint16m4_t vnclip(vint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16m4_m))) +vint16m4_t vnclip(vbool4_t op0, vint16m4_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16mf2))) +vint16mf2_t vnclip(vint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16mf2_m))) +vint16mf2_t vnclip(vbool32_t op0, vint16mf2_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16mf4))) +vint16mf4_t vnclip(vint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16mf4_m))) +vint16mf4_t vnclip(vbool64_t op0, vint16mf4_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32m1))) +vint32m1_t vnclip(vint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32m1_m))) +vint32m1_t vnclip(vbool32_t op0, vint32m1_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32m2))) +vint32m2_t vnclip(vint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32m2_m))) +vint32m2_t vnclip(vbool16_t op0, vint32m2_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32m4))) +vint32m4_t vnclip(vint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32m4_m))) +vint32m4_t vnclip(vbool8_t op0, vint32m4_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32mf2))) +vint32mf2_t vnclip(vint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32mf2_m))) +vint32mf2_t vnclip(vbool64_t op0, vint32mf2_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m1))) +void vsuxei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m1_m))) +void vsuxei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m2))) +void vsuxei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m2_m))) +void vsuxei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m4))) +void vsuxei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m4_m))) +void vsuxei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m8))) +void vsuxei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m8_m))) +void vsuxei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m1))) +void vsuxei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m1_m))) +void vsuxei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m2))) +void vsuxei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m2_m))) +void vsuxei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m4))) +void vsuxei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m4_m))) +void vsuxei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m8))) +void vsuxei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m8_m))) +void vsuxei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m1))) +void vsuxei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m1_m))) +void vsuxei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m2))) +void vsuxei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m2_m))) +void vsuxei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m4))) +void vsuxei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m4_m))) +void vsuxei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m8))) +void vsuxei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m8_m))) +void vsuxei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m1))) +void vsuxei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m1_m))) +void vsuxei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m2))) +void vsuxei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m2_m))) +void vsuxei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m4))) +void vsuxei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m4_m))) +void vsuxei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m8))) +void vsuxei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m8_m))) +void vsuxei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m1))) +void vsuxei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m1_m))) +void vsuxei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m2))) +void vsuxei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m2_m))) +void vsuxei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m4))) +void vsuxei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m4_m))) +void vsuxei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m8))) +void vsuxei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m8_m))) +void vsuxei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m1))) +void vsuxei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m1_m))) +void vsuxei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m2))) +void vsuxei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m2_m))) +void vsuxei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m4))) +void vsuxei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m4_m))) +void vsuxei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m8))) +void vsuxei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m8_m))) +void vsuxei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m1))) +void vsuxei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m1_m))) +void vsuxei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m2))) +void vsuxei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m2_m))) +void vsuxei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m4))) +void vsuxei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m4_m))) +void vsuxei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m8))) +void vsuxei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m8_m))) +void vsuxei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m1_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m1_i8m1_m))) +vint8m1_t vredsum(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m2_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m2_i8m1_m))) +vint8m1_t vredsum(vbool4_t op0, vint8m1_t op1, vint8m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m4_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m4_i8m1_m))) +vint8m1_t vredsum(vbool2_t op0, vint8m1_t op1, vint8m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m8_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m8_i8m1_m))) +vint8m1_t vredsum(vbool1_t op0, vint8m1_t op1, vint8m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8mf2_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8mf2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8mf2_i8m1_m))) +vint8m1_t vredsum(vbool16_t op0, vint8m1_t op1, vint8mf2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8mf4_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8mf4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8mf4_i8m1_m))) +vint8m1_t vredsum(vbool32_t op0, vint8m1_t op1, vint8mf4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8mf8_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8mf8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8mf8_i8m1_m))) +vint8m1_t vredsum(vbool64_t op0, vint8m1_t op1, vint8mf8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m1_i16m1))) +vint16m1_t vredsum(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m1_i16m1_m))) +vint16m1_t vredsum(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m2_i16m1))) +vint16m1_t vredsum(vint16m1_t op0, vint16m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m2_i16m1_m))) +vint16m1_t vredsum(vbool8_t op0, vint16m1_t op1, vint16m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m4_i16m1))) +vint16m1_t vredsum(vint16m1_t op0, vint16m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m4_i16m1_m))) +vint16m1_t vredsum(vbool4_t op0, vint16m1_t op1, vint16m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m8_i16m1))) +vint16m1_t vredsum(vint16m1_t op0, vint16m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m8_i16m1_m))) +vint16m1_t vredsum(vbool2_t op0, vint16m1_t op1, vint16m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16mf2_i16m1))) +vint16m1_t vredsum(vint16m1_t op0, vint16mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16mf2_i16m1_m))) +vint16m1_t vredsum(vbool32_t op0, vint16m1_t op1, vint16mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16mf4_i16m1))) +vint16m1_t vredsum(vint16m1_t op0, vint16mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16mf4_i16m1_m))) +vint16m1_t vredsum(vbool64_t op0, vint16m1_t op1, vint16mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m1_i32m1))) +vint32m1_t vredsum(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m1_i32m1_m))) +vint32m1_t vredsum(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m2_i32m1))) +vint32m1_t vredsum(vint32m1_t op0, vint32m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m2_i32m1_m))) +vint32m1_t vredsum(vbool16_t op0, vint32m1_t op1, vint32m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m4_i32m1))) +vint32m1_t vredsum(vint32m1_t op0, vint32m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m4_i32m1_m))) +vint32m1_t vredsum(vbool8_t op0, vint32m1_t op1, vint32m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m8_i32m1))) +vint32m1_t vredsum(vint32m1_t op0, vint32m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m8_i32m1_m))) +vint32m1_t vredsum(vbool4_t op0, vint32m1_t op1, vint32m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32mf2_i32m1))) +vint32m1_t vredsum(vint32m1_t op0, vint32mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32mf2_i32m1_m))) +vint32m1_t vredsum(vbool64_t op0, vint32m1_t op1, vint32mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m1_i64m1))) +vint64m1_t vredsum(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m1_i64m1_m))) +vint64m1_t vredsum(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m2_i64m1))) +vint64m1_t vredsum(vint64m1_t op0, vint64m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m2_i64m1_m))) +vint64m1_t vredsum(vbool32_t op0, vint64m1_t op1, vint64m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m4_i64m1))) +vint64m1_t vredsum(vint64m1_t op0, vint64m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m4_i64m1_m))) +vint64m1_t vredsum(vbool16_t op0, vint64m1_t op1, vint64m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m8_i64m1))) +vint64m1_t vredsum(vint64m1_t op0, vint64m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m8_i64m1_m))) +vint64m1_t vredsum(vbool8_t op0, vint64m1_t op1, vint64m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m1_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m1_u8m1_m))) +vuint8m1_t vredsum(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m2_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m2_u8m1_m))) +vuint8m1_t vredsum(vbool4_t op0, vuint8m1_t op1, vuint8m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m4_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m4_u8m1_m))) +vuint8m1_t vredsum(vbool2_t op0, vuint8m1_t op1, vuint8m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m8_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m8_u8m1_m))) +vuint8m1_t vredsum(vbool1_t op0, vuint8m1_t op1, vuint8m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8mf2_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8mf2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8mf2_u8m1_m))) +vuint8m1_t vredsum(vbool16_t op0, vuint8m1_t op1, vuint8mf2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8mf4_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8mf4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8mf4_u8m1_m))) +vuint8m1_t vredsum(vbool32_t op0, vuint8m1_t op1, vuint8mf4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8mf8_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8mf8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8mf8_u8m1_m))) +vuint8m1_t vredsum(vbool64_t op0, vuint8m1_t op1, vuint8mf8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m1_u16m1))) +vuint16m1_t vredsum(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m1_u16m1_m))) +vuint16m1_t vredsum(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m2_u16m1))) +vuint16m1_t vredsum(vuint16m1_t op0, vuint16m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m2_u16m1_m))) +vuint16m1_t vredsum(vbool8_t op0, vuint16m1_t op1, vuint16m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m4_u16m1))) +vuint16m1_t vredsum(vuint16m1_t op0, vuint16m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m4_u16m1_m))) +vuint16m1_t vredsum(vbool4_t op0, vuint16m1_t op1, vuint16m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m8_u16m1))) +vuint16m1_t vredsum(vuint16m1_t op0, vuint16m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m8_u16m1_m))) +vuint16m1_t vredsum(vbool2_t op0, vuint16m1_t op1, vuint16m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16mf2_u16m1))) +vuint16m1_t vredsum(vuint16m1_t op0, vuint16mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16mf2_u16m1_m))) +vuint16m1_t vredsum(vbool32_t op0, vuint16m1_t op1, vuint16mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16mf4_u16m1))) +vuint16m1_t vredsum(vuint16m1_t op0, vuint16mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16mf4_u16m1_m))) +vuint16m1_t vredsum(vbool64_t op0, vuint16m1_t op1, vuint16mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m1_u32m1))) +vuint32m1_t vredsum(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m1_u32m1_m))) +vuint32m1_t vredsum(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m2_u32m1))) +vuint32m1_t vredsum(vuint32m1_t op0, vuint32m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m2_u32m1_m))) +vuint32m1_t vredsum(vbool16_t op0, vuint32m1_t op1, vuint32m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m4_u32m1))) +vuint32m1_t vredsum(vuint32m1_t op0, vuint32m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m4_u32m1_m))) +vuint32m1_t vredsum(vbool8_t op0, vuint32m1_t op1, vuint32m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m8_u32m1))) +vuint32m1_t vredsum(vuint32m1_t op0, vuint32m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m8_u32m1_m))) +vuint32m1_t vredsum(vbool4_t op0, vuint32m1_t op1, vuint32m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32mf2_u32m1))) +vuint32m1_t vredsum(vuint32m1_t op0, vuint32mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32mf2_u32m1_m))) +vuint32m1_t vredsum(vbool64_t op0, vuint32m1_t op1, vuint32mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m1_u64m1))) +vuint64m1_t vredsum(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m1_u64m1_m))) +vuint64m1_t vredsum(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m2_u64m1))) +vuint64m1_t vredsum(vuint64m1_t op0, vuint64m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m2_u64m1_m))) +vuint64m1_t vredsum(vbool32_t op0, vuint64m1_t op1, vuint64m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m4_u64m1))) +vuint64m1_t vredsum(vuint64m1_t op0, vuint64m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m4_u64m1_m))) +vuint64m1_t vredsum(vbool16_t op0, vuint64m1_t op1, vuint64m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m8_u64m1))) +vuint64m1_t vredsum(vuint64m1_t op0, vuint64m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m8_u64m1_m))) +vuint64m1_t vredsum(vbool8_t op0, vuint64m1_t op1, vuint64m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m1_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m1_u8m1_m))) +vuint8m1_t vredmaxu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m2_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m2_u8m1_m))) +vuint8m1_t vredmaxu(vbool4_t op0, vuint8m1_t op1, vuint8m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m4_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m4_u8m1_m))) +vuint8m1_t vredmaxu(vbool2_t op0, vuint8m1_t op1, vuint8m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m8_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m8_u8m1_m))) +vuint8m1_t vredmaxu(vbool1_t op0, vuint8m1_t op1, vuint8m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8mf2_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8mf2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8mf2_u8m1_m))) +vuint8m1_t vredmaxu(vbool16_t op0, vuint8m1_t op1, vuint8mf2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8mf4_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8mf4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8mf4_u8m1_m))) +vuint8m1_t vredmaxu(vbool32_t op0, vuint8m1_t op1, vuint8mf4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8mf8_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8mf8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8mf8_u8m1_m))) +vuint8m1_t vredmaxu(vbool64_t op0, vuint8m1_t op1, vuint8mf8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m1_u16m1))) +vuint16m1_t vredmaxu(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m1_u16m1_m))) +vuint16m1_t vredmaxu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m2_u16m1))) +vuint16m1_t vredmaxu(vuint16m1_t op0, vuint16m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m2_u16m1_m))) +vuint16m1_t vredmaxu(vbool8_t op0, vuint16m1_t op1, vuint16m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m4_u16m1))) +vuint16m1_t vredmaxu(vuint16m1_t op0, vuint16m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m4_u16m1_m))) +vuint16m1_t vredmaxu(vbool4_t op0, vuint16m1_t op1, vuint16m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m8_u16m1))) +vuint16m1_t vredmaxu(vuint16m1_t op0, vuint16m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m8_u16m1_m))) +vuint16m1_t vredmaxu(vbool2_t op0, vuint16m1_t op1, vuint16m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16mf2_u16m1))) +vuint16m1_t vredmaxu(vuint16m1_t op0, vuint16mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16mf2_u16m1_m))) +vuint16m1_t vredmaxu(vbool32_t op0, vuint16m1_t op1, vuint16mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16mf4_u16m1))) +vuint16m1_t vredmaxu(vuint16m1_t op0, vuint16mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16mf4_u16m1_m))) +vuint16m1_t vredmaxu(vbool64_t op0, vuint16m1_t op1, vuint16mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m1_u32m1))) +vuint32m1_t vredmaxu(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m1_u32m1_m))) +vuint32m1_t vredmaxu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m2_u32m1))) +vuint32m1_t vredmaxu(vuint32m1_t op0, vuint32m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m2_u32m1_m))) +vuint32m1_t vredmaxu(vbool16_t op0, vuint32m1_t op1, vuint32m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m4_u32m1))) +vuint32m1_t vredmaxu(vuint32m1_t op0, vuint32m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m4_u32m1_m))) +vuint32m1_t vredmaxu(vbool8_t op0, vuint32m1_t op1, vuint32m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m8_u32m1))) +vuint32m1_t vredmaxu(vuint32m1_t op0, vuint32m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m8_u32m1_m))) +vuint32m1_t vredmaxu(vbool4_t op0, vuint32m1_t op1, vuint32m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32mf2_u32m1))) +vuint32m1_t vredmaxu(vuint32m1_t op0, vuint32mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32mf2_u32m1_m))) +vuint32m1_t vredmaxu(vbool64_t op0, vuint32m1_t op1, vuint32mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m1_u64m1))) +vuint64m1_t vredmaxu(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m1_u64m1_m))) +vuint64m1_t vredmaxu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m2_u64m1))) +vuint64m1_t vredmaxu(vuint64m1_t op0, vuint64m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m2_u64m1_m))) +vuint64m1_t vredmaxu(vbool32_t op0, vuint64m1_t op1, vuint64m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m4_u64m1))) +vuint64m1_t vredmaxu(vuint64m1_t op0, vuint64m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m4_u64m1_m))) +vuint64m1_t vredmaxu(vbool16_t op0, vuint64m1_t op1, vuint64m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m8_u64m1))) +vuint64m1_t vredmaxu(vuint64m1_t op0, vuint64m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m8_u64m1_m))) +vuint64m1_t vredmaxu(vbool8_t op0, vuint64m1_t op1, vuint64m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m1_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m1_i8m1_m))) +vint8m1_t vredmax(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m2_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m2_i8m1_m))) +vint8m1_t vredmax(vbool4_t op0, vint8m1_t op1, vint8m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m4_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m4_i8m1_m))) +vint8m1_t vredmax(vbool2_t op0, vint8m1_t op1, vint8m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m8_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m8_i8m1_m))) +vint8m1_t vredmax(vbool1_t op0, vint8m1_t op1, vint8m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8mf2_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8mf2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8mf2_i8m1_m))) +vint8m1_t vredmax(vbool16_t op0, vint8m1_t op1, vint8mf2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8mf4_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8mf4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8mf4_i8m1_m))) +vint8m1_t vredmax(vbool32_t op0, vint8m1_t op1, vint8mf4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8mf8_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8mf8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8mf8_i8m1_m))) +vint8m1_t vredmax(vbool64_t op0, vint8m1_t op1, vint8mf8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m1_i16m1))) +vint16m1_t vredmax(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m1_i16m1_m))) +vint16m1_t vredmax(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m2_i16m1))) +vint16m1_t vredmax(vint16m1_t op0, vint16m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m2_i16m1_m))) +vint16m1_t vredmax(vbool8_t op0, vint16m1_t op1, vint16m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m4_i16m1))) +vint16m1_t vredmax(vint16m1_t op0, vint16m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m4_i16m1_m))) +vint16m1_t vredmax(vbool4_t op0, vint16m1_t op1, vint16m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m8_i16m1))) +vint16m1_t vredmax(vint16m1_t op0, vint16m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m8_i16m1_m))) +vint16m1_t vredmax(vbool2_t op0, vint16m1_t op1, vint16m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16mf2_i16m1))) +vint16m1_t vredmax(vint16m1_t op0, vint16mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16mf2_i16m1_m))) +vint16m1_t vredmax(vbool32_t op0, vint16m1_t op1, vint16mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16mf4_i16m1))) +vint16m1_t vredmax(vint16m1_t op0, vint16mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16mf4_i16m1_m))) +vint16m1_t vredmax(vbool64_t op0, vint16m1_t op1, vint16mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m1_i32m1))) +vint32m1_t vredmax(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m1_i32m1_m))) +vint32m1_t vredmax(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m2_i32m1))) +vint32m1_t vredmax(vint32m1_t op0, vint32m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m2_i32m1_m))) +vint32m1_t vredmax(vbool16_t op0, vint32m1_t op1, vint32m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m4_i32m1))) +vint32m1_t vredmax(vint32m1_t op0, vint32m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m4_i32m1_m))) +vint32m1_t vredmax(vbool8_t op0, vint32m1_t op1, vint32m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m8_i32m1))) +vint32m1_t vredmax(vint32m1_t op0, vint32m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m8_i32m1_m))) +vint32m1_t vredmax(vbool4_t op0, vint32m1_t op1, vint32m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32mf2_i32m1))) +vint32m1_t vredmax(vint32m1_t op0, vint32mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32mf2_i32m1_m))) +vint32m1_t vredmax(vbool64_t op0, vint32m1_t op1, vint32mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m1_i64m1))) +vint64m1_t vredmax(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m1_i64m1_m))) +vint64m1_t vredmax(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m2_i64m1))) +vint64m1_t vredmax(vint64m1_t op0, vint64m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m2_i64m1_m))) +vint64m1_t vredmax(vbool32_t op0, vint64m1_t op1, vint64m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m4_i64m1))) +vint64m1_t vredmax(vint64m1_t op0, vint64m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m4_i64m1_m))) +vint64m1_t vredmax(vbool16_t op0, vint64m1_t op1, vint64m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m8_i64m1))) +vint64m1_t vredmax(vint64m1_t op0, vint64m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m8_i64m1_m))) +vint64m1_t vredmax(vbool8_t op0, vint64m1_t op1, vint64m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m1_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m1_u8m1_m))) +vuint8m1_t vredminu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m2_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m2_u8m1_m))) +vuint8m1_t vredminu(vbool4_t op0, vuint8m1_t op1, vuint8m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m4_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m4_u8m1_m))) +vuint8m1_t vredminu(vbool2_t op0, vuint8m1_t op1, vuint8m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m8_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m8_u8m1_m))) +vuint8m1_t vredminu(vbool1_t op0, vuint8m1_t op1, vuint8m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8mf2_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8mf2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8mf2_u8m1_m))) +vuint8m1_t vredminu(vbool16_t op0, vuint8m1_t op1, vuint8mf2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8mf4_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8mf4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8mf4_u8m1_m))) +vuint8m1_t vredminu(vbool32_t op0, vuint8m1_t op1, vuint8mf4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8mf8_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8mf8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8mf8_u8m1_m))) +vuint8m1_t vredminu(vbool64_t op0, vuint8m1_t op1, vuint8mf8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m1_u16m1))) +vuint16m1_t vredminu(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m1_u16m1_m))) +vuint16m1_t vredminu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m2_u16m1))) +vuint16m1_t vredminu(vuint16m1_t op0, vuint16m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m2_u16m1_m))) +vuint16m1_t vredminu(vbool8_t op0, vuint16m1_t op1, vuint16m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m4_u16m1))) +vuint16m1_t vredminu(vuint16m1_t op0, vuint16m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m4_u16m1_m))) +vuint16m1_t vredminu(vbool4_t op0, vuint16m1_t op1, vuint16m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m8_u16m1))) +vuint16m1_t vredminu(vuint16m1_t op0, vuint16m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m8_u16m1_m))) +vuint16m1_t vredminu(vbool2_t op0, vuint16m1_t op1, vuint16m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16mf2_u16m1))) +vuint16m1_t vredminu(vuint16m1_t op0, vuint16mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16mf2_u16m1_m))) +vuint16m1_t vredminu(vbool32_t op0, vuint16m1_t op1, vuint16mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16mf4_u16m1))) +vuint16m1_t vredminu(vuint16m1_t op0, vuint16mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16mf4_u16m1_m))) +vuint16m1_t vredminu(vbool64_t op0, vuint16m1_t op1, vuint16mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m1_u32m1))) +vuint32m1_t vredminu(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m1_u32m1_m))) +vuint32m1_t vredminu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m2_u32m1))) +vuint32m1_t vredminu(vuint32m1_t op0, vuint32m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m2_u32m1_m))) +vuint32m1_t vredminu(vbool16_t op0, vuint32m1_t op1, vuint32m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m4_u32m1))) +vuint32m1_t vredminu(vuint32m1_t op0, vuint32m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m4_u32m1_m))) +vuint32m1_t vredminu(vbool8_t op0, vuint32m1_t op1, vuint32m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m8_u32m1))) +vuint32m1_t vredminu(vuint32m1_t op0, vuint32m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m8_u32m1_m))) +vuint32m1_t vredminu(vbool4_t op0, vuint32m1_t op1, vuint32m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32mf2_u32m1))) +vuint32m1_t vredminu(vuint32m1_t op0, vuint32mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32mf2_u32m1_m))) +vuint32m1_t vredminu(vbool64_t op0, vuint32m1_t op1, vuint32mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m1_u64m1))) +vuint64m1_t vredminu(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m1_u64m1_m))) +vuint64m1_t vredminu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m2_u64m1))) +vuint64m1_t vredminu(vuint64m1_t op0, vuint64m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m2_u64m1_m))) +vuint64m1_t vredminu(vbool32_t op0, vuint64m1_t op1, vuint64m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m4_u64m1))) +vuint64m1_t vredminu(vuint64m1_t op0, vuint64m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m4_u64m1_m))) +vuint64m1_t vredminu(vbool16_t op0, vuint64m1_t op1, vuint64m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m8_u64m1))) +vuint64m1_t vredminu(vuint64m1_t op0, vuint64m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m8_u64m1_m))) +vuint64m1_t vredminu(vbool8_t op0, vuint64m1_t op1, vuint64m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m1_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m1_i8m1_m))) +vint8m1_t vredmin(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m2_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m2_i8m1_m))) +vint8m1_t vredmin(vbool4_t op0, vint8m1_t op1, vint8m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m4_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m4_i8m1_m))) +vint8m1_t vredmin(vbool2_t op0, vint8m1_t op1, vint8m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m8_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m8_i8m1_m))) +vint8m1_t vredmin(vbool1_t op0, vint8m1_t op1, vint8m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8mf2_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8mf2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8mf2_i8m1_m))) +vint8m1_t vredmin(vbool16_t op0, vint8m1_t op1, vint8mf2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8mf4_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8mf4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8mf4_i8m1_m))) +vint8m1_t vredmin(vbool32_t op0, vint8m1_t op1, vint8mf4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8mf8_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8mf8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8mf8_i8m1_m))) +vint8m1_t vredmin(vbool64_t op0, vint8m1_t op1, vint8mf8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m1_i16m1))) +vint16m1_t vredmin(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m1_i16m1_m))) +vint16m1_t vredmin(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m2_i16m1))) +vint16m1_t vredmin(vint16m1_t op0, vint16m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m2_i16m1_m))) +vint16m1_t vredmin(vbool8_t op0, vint16m1_t op1, vint16m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m4_i16m1))) +vint16m1_t vredmin(vint16m1_t op0, vint16m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m4_i16m1_m))) +vint16m1_t vredmin(vbool4_t op0, vint16m1_t op1, vint16m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m8_i16m1))) +vint16m1_t vredmin(vint16m1_t op0, vint16m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m8_i16m1_m))) +vint16m1_t vredmin(vbool2_t op0, vint16m1_t op1, vint16m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16mf2_i16m1))) +vint16m1_t vredmin(vint16m1_t op0, vint16mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16mf2_i16m1_m))) +vint16m1_t vredmin(vbool32_t op0, vint16m1_t op1, vint16mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16mf4_i16m1))) +vint16m1_t vredmin(vint16m1_t op0, vint16mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16mf4_i16m1_m))) +vint16m1_t vredmin(vbool64_t op0, vint16m1_t op1, vint16mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m1_i32m1))) +vint32m1_t vredmin(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m1_i32m1_m))) +vint32m1_t vredmin(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m2_i32m1))) +vint32m1_t vredmin(vint32m1_t op0, vint32m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m2_i32m1_m))) +vint32m1_t vredmin(vbool16_t op0, vint32m1_t op1, vint32m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m4_i32m1))) +vint32m1_t vredmin(vint32m1_t op0, vint32m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m4_i32m1_m))) +vint32m1_t vredmin(vbool8_t op0, vint32m1_t op1, vint32m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m8_i32m1))) +vint32m1_t vredmin(vint32m1_t op0, vint32m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m8_i32m1_m))) +vint32m1_t vredmin(vbool4_t op0, vint32m1_t op1, vint32m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32mf2_i32m1))) +vint32m1_t vredmin(vint32m1_t op0, vint32mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32mf2_i32m1_m))) +vint32m1_t vredmin(vbool64_t op0, vint32m1_t op1, vint32mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m1_i64m1))) +vint64m1_t vredmin(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m1_i64m1_m))) +vint64m1_t vredmin(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m2_i64m1))) +vint64m1_t vredmin(vint64m1_t op0, vint64m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m2_i64m1_m))) +vint64m1_t vredmin(vbool32_t op0, vint64m1_t op1, vint64m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m4_i64m1))) +vint64m1_t vredmin(vint64m1_t op0, vint64m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m4_i64m1_m))) +vint64m1_t vredmin(vbool16_t op0, vint64m1_t op1, vint64m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m8_i64m1))) +vint64m1_t vredmin(vint64m1_t op0, vint64m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m8_i64m1_m))) +vint64m1_t vredmin(vbool8_t op0, vint64m1_t op1, vint64m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m1_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m1_i8m1_m))) +vint8m1_t vredand(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m2_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m2_i8m1_m))) +vint8m1_t vredand(vbool4_t op0, vint8m1_t op1, vint8m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m4_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m4_i8m1_m))) +vint8m1_t vredand(vbool2_t op0, vint8m1_t op1, vint8m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m8_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m8_i8m1_m))) +vint8m1_t vredand(vbool1_t op0, vint8m1_t op1, vint8m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8mf2_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8mf2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8mf2_i8m1_m))) +vint8m1_t vredand(vbool16_t op0, vint8m1_t op1, vint8mf2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8mf4_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8mf4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8mf4_i8m1_m))) +vint8m1_t vredand(vbool32_t op0, vint8m1_t op1, vint8mf4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8mf8_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8mf8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8mf8_i8m1_m))) +vint8m1_t vredand(vbool64_t op0, vint8m1_t op1, vint8mf8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m1_i16m1))) +vint16m1_t vredand(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m1_i16m1_m))) +vint16m1_t vredand(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m2_i16m1))) +vint16m1_t vredand(vint16m1_t op0, vint16m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m2_i16m1_m))) +vint16m1_t vredand(vbool8_t op0, vint16m1_t op1, vint16m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m4_i16m1))) +vint16m1_t vredand(vint16m1_t op0, vint16m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m4_i16m1_m))) +vint16m1_t vredand(vbool4_t op0, vint16m1_t op1, vint16m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m8_i16m1))) +vint16m1_t vredand(vint16m1_t op0, vint16m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m8_i16m1_m))) +vint16m1_t vredand(vbool2_t op0, vint16m1_t op1, vint16m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16mf2_i16m1))) +vint16m1_t vredand(vint16m1_t op0, vint16mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16mf2_i16m1_m))) +vint16m1_t vredand(vbool32_t op0, vint16m1_t op1, vint16mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16mf4_i16m1))) +vint16m1_t vredand(vint16m1_t op0, vint16mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16mf4_i16m1_m))) +vint16m1_t vredand(vbool64_t op0, vint16m1_t op1, vint16mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m1_i32m1))) +vint32m1_t vredand(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m1_i32m1_m))) +vint32m1_t vredand(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m2_i32m1))) +vint32m1_t vredand(vint32m1_t op0, vint32m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m2_i32m1_m))) +vint32m1_t vredand(vbool16_t op0, vint32m1_t op1, vint32m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m4_i32m1))) +vint32m1_t vredand(vint32m1_t op0, vint32m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m4_i32m1_m))) +vint32m1_t vredand(vbool8_t op0, vint32m1_t op1, vint32m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m8_i32m1))) +vint32m1_t vredand(vint32m1_t op0, vint32m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m8_i32m1_m))) +vint32m1_t vredand(vbool4_t op0, vint32m1_t op1, vint32m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32mf2_i32m1))) +vint32m1_t vredand(vint32m1_t op0, vint32mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32mf2_i32m1_m))) +vint32m1_t vredand(vbool64_t op0, vint32m1_t op1, vint32mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m1_i64m1))) +vint64m1_t vredand(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m1_i64m1_m))) +vint64m1_t vredand(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m2_i64m1))) +vint64m1_t vredand(vint64m1_t op0, vint64m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m2_i64m1_m))) +vint64m1_t vredand(vbool32_t op0, vint64m1_t op1, vint64m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m4_i64m1))) +vint64m1_t vredand(vint64m1_t op0, vint64m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m4_i64m1_m))) +vint64m1_t vredand(vbool16_t op0, vint64m1_t op1, vint64m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m8_i64m1))) +vint64m1_t vredand(vint64m1_t op0, vint64m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m8_i64m1_m))) +vint64m1_t vredand(vbool8_t op0, vint64m1_t op1, vint64m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m1_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m1_u8m1_m))) +vuint8m1_t vredand(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m2_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m2_u8m1_m))) +vuint8m1_t vredand(vbool4_t op0, vuint8m1_t op1, vuint8m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m4_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m4_u8m1_m))) +vuint8m1_t vredand(vbool2_t op0, vuint8m1_t op1, vuint8m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m8_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m8_u8m1_m))) +vuint8m1_t vredand(vbool1_t op0, vuint8m1_t op1, vuint8m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8mf2_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8mf2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8mf2_u8m1_m))) +vuint8m1_t vredand(vbool16_t op0, vuint8m1_t op1, vuint8mf2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8mf4_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8mf4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8mf4_u8m1_m))) +vuint8m1_t vredand(vbool32_t op0, vuint8m1_t op1, vuint8mf4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8mf8_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8mf8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8mf8_u8m1_m))) +vuint8m1_t vredand(vbool64_t op0, vuint8m1_t op1, vuint8mf8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m1_u16m1))) +vuint16m1_t vredand(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m1_u16m1_m))) +vuint16m1_t vredand(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m2_u16m1))) +vuint16m1_t vredand(vuint16m1_t op0, vuint16m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m2_u16m1_m))) +vuint16m1_t vredand(vbool8_t op0, vuint16m1_t op1, vuint16m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m4_u16m1))) +vuint16m1_t vredand(vuint16m1_t op0, vuint16m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m4_u16m1_m))) +vuint16m1_t vredand(vbool4_t op0, vuint16m1_t op1, vuint16m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m8_u16m1))) +vuint16m1_t vredand(vuint16m1_t op0, vuint16m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m8_u16m1_m))) +vuint16m1_t vredand(vbool2_t op0, vuint16m1_t op1, vuint16m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16mf2_u16m1))) +vuint16m1_t vredand(vuint16m1_t op0, vuint16mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16mf2_u16m1_m))) +vuint16m1_t vredand(vbool32_t op0, vuint16m1_t op1, vuint16mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16mf4_u16m1))) +vuint16m1_t vredand(vuint16m1_t op0, vuint16mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16mf4_u16m1_m))) +vuint16m1_t vredand(vbool64_t op0, vuint16m1_t op1, vuint16mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m1_u32m1))) +vuint32m1_t vredand(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m1_u32m1_m))) +vuint32m1_t vredand(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m2_u32m1))) +vuint32m1_t vredand(vuint32m1_t op0, vuint32m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m2_u32m1_m))) +vuint32m1_t vredand(vbool16_t op0, vuint32m1_t op1, vuint32m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m4_u32m1))) +vuint32m1_t vredand(vuint32m1_t op0, vuint32m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m4_u32m1_m))) +vuint32m1_t vredand(vbool8_t op0, vuint32m1_t op1, vuint32m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m8_u32m1))) +vuint32m1_t vredand(vuint32m1_t op0, vuint32m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m8_u32m1_m))) +vuint32m1_t vredand(vbool4_t op0, vuint32m1_t op1, vuint32m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32mf2_u32m1))) +vuint32m1_t vredand(vuint32m1_t op0, vuint32mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32mf2_u32m1_m))) +vuint32m1_t vredand(vbool64_t op0, vuint32m1_t op1, vuint32mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m1_u64m1))) +vuint64m1_t vredand(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m1_u64m1_m))) +vuint64m1_t vredand(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m2_u64m1))) +vuint64m1_t vredand(vuint64m1_t op0, vuint64m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m2_u64m1_m))) +vuint64m1_t vredand(vbool32_t op0, vuint64m1_t op1, vuint64m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m4_u64m1))) +vuint64m1_t vredand(vuint64m1_t op0, vuint64m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m4_u64m1_m))) +vuint64m1_t vredand(vbool16_t op0, vuint64m1_t op1, vuint64m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m8_u64m1))) +vuint64m1_t vredand(vuint64m1_t op0, vuint64m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m8_u64m1_m))) +vuint64m1_t vredand(vbool8_t op0, vuint64m1_t op1, vuint64m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m1))) +void vsuxei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m1_m))) +void vsuxei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m2))) +void vsuxei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m2_m))) +void vsuxei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m4))) +void vsuxei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m4_m))) +void vsuxei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m8))) +void vsuxei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m8_m))) +void vsuxei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m1_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m1_i8m1_m))) +vint8m1_t vredor(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m2_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m2_i8m1_m))) +vint8m1_t vredor(vbool4_t op0, vint8m1_t op1, vint8m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m4_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m4_i8m1_m))) +vint8m1_t vredor(vbool2_t op0, vint8m1_t op1, vint8m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m8_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m8_i8m1_m))) +vint8m1_t vredor(vbool1_t op0, vint8m1_t op1, vint8m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8mf2_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8mf2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8mf2_i8m1_m))) +vint8m1_t vredor(vbool16_t op0, vint8m1_t op1, vint8mf2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8mf4_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8mf4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8mf4_i8m1_m))) +vint8m1_t vredor(vbool32_t op0, vint8m1_t op1, vint8mf4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8mf8_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8mf8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8mf8_i8m1_m))) +vint8m1_t vredor(vbool64_t op0, vint8m1_t op1, vint8mf8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m1_i16m1))) +vint16m1_t vredor(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m1_i16m1_m))) +vint16m1_t vredor(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m2_i16m1))) +vint16m1_t vredor(vint16m1_t op0, vint16m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m2_i16m1_m))) +vint16m1_t vredor(vbool8_t op0, vint16m1_t op1, vint16m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m4_i16m1))) +vint16m1_t vredor(vint16m1_t op0, vint16m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m4_i16m1_m))) +vint16m1_t vredor(vbool4_t op0, vint16m1_t op1, vint16m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m8_i16m1))) +vint16m1_t vredor(vint16m1_t op0, vint16m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m8_i16m1_m))) +vint16m1_t vredor(vbool2_t op0, vint16m1_t op1, vint16m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16mf2_i16m1))) +vint16m1_t vredor(vint16m1_t op0, vint16mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16mf2_i16m1_m))) +vint16m1_t vredor(vbool32_t op0, vint16m1_t op1, vint16mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16mf4_i16m1))) +vint16m1_t vredor(vint16m1_t op0, vint16mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16mf4_i16m1_m))) +vint16m1_t vredor(vbool64_t op0, vint16m1_t op1, vint16mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m1_i32m1))) +vint32m1_t vredor(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m1_i32m1_m))) +vint32m1_t vredor(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m2_i32m1))) +vint32m1_t vredor(vint32m1_t op0, vint32m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m2_i32m1_m))) +vint32m1_t vredor(vbool16_t op0, vint32m1_t op1, vint32m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m4_i32m1))) +vint32m1_t vredor(vint32m1_t op0, vint32m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m4_i32m1_m))) +vint32m1_t vredor(vbool8_t op0, vint32m1_t op1, vint32m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m8_i32m1))) +vint32m1_t vredor(vint32m1_t op0, vint32m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m8_i32m1_m))) +vint32m1_t vredor(vbool4_t op0, vint32m1_t op1, vint32m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32mf2_i32m1))) +vint32m1_t vredor(vint32m1_t op0, vint32mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32mf2_i32m1_m))) +vint32m1_t vredor(vbool64_t op0, vint32m1_t op1, vint32mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m1_i64m1))) +vint64m1_t vredor(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m1_i64m1_m))) +vint64m1_t vredor(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m2_i64m1))) +vint64m1_t vredor(vint64m1_t op0, vint64m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m2_i64m1_m))) +vint64m1_t vredor(vbool32_t op0, vint64m1_t op1, vint64m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m4_i64m1))) +vint64m1_t vredor(vint64m1_t op0, vint64m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m4_i64m1_m))) +vint64m1_t vredor(vbool16_t op0, vint64m1_t op1, vint64m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m8_i64m1))) +vint64m1_t vredor(vint64m1_t op0, vint64m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m8_i64m1_m))) +vint64m1_t vredor(vbool8_t op0, vint64m1_t op1, vint64m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m1_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m1_u8m1_m))) +vuint8m1_t vredor(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m2_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m2_u8m1_m))) +vuint8m1_t vredor(vbool4_t op0, vuint8m1_t op1, vuint8m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m4_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m4_u8m1_m))) +vuint8m1_t vredor(vbool2_t op0, vuint8m1_t op1, vuint8m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m8_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m8_u8m1_m))) +vuint8m1_t vredor(vbool1_t op0, vuint8m1_t op1, vuint8m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8mf2_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8mf2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8mf2_u8m1_m))) +vuint8m1_t vredor(vbool16_t op0, vuint8m1_t op1, vuint8mf2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8mf4_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8mf4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8mf4_u8m1_m))) +vuint8m1_t vredor(vbool32_t op0, vuint8m1_t op1, vuint8mf4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8mf8_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8mf8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8mf8_u8m1_m))) +vuint8m1_t vredor(vbool64_t op0, vuint8m1_t op1, vuint8mf8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m1_u16m1))) +vuint16m1_t vredor(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m1_u16m1_m))) +vuint16m1_t vredor(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m2_u16m1))) +vuint16m1_t vredor(vuint16m1_t op0, vuint16m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m2_u16m1_m))) +vuint16m1_t vredor(vbool8_t op0, vuint16m1_t op1, vuint16m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m4_u16m1))) +vuint16m1_t vredor(vuint16m1_t op0, vuint16m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m4_u16m1_m))) +vuint16m1_t vredor(vbool4_t op0, vuint16m1_t op1, vuint16m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m8_u16m1))) +vuint16m1_t vredor(vuint16m1_t op0, vuint16m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m8_u16m1_m))) +vuint16m1_t vredor(vbool2_t op0, vuint16m1_t op1, vuint16m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16mf2_u16m1))) +vuint16m1_t vredor(vuint16m1_t op0, vuint16mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16mf2_u16m1_m))) +vuint16m1_t vredor(vbool32_t op0, vuint16m1_t op1, vuint16mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16mf4_u16m1))) +vuint16m1_t vredor(vuint16m1_t op0, vuint16mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16mf4_u16m1_m))) +vuint16m1_t vredor(vbool64_t op0, vuint16m1_t op1, vuint16mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m1_u32m1))) +vuint32m1_t vredor(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m1_u32m1_m))) +vuint32m1_t vredor(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m2_u32m1))) +vuint32m1_t vredor(vuint32m1_t op0, vuint32m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m2_u32m1_m))) +vuint32m1_t vredor(vbool16_t op0, vuint32m1_t op1, vuint32m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m4_u32m1))) +vuint32m1_t vredor(vuint32m1_t op0, vuint32m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m4_u32m1_m))) +vuint32m1_t vredor(vbool8_t op0, vuint32m1_t op1, vuint32m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m8_u32m1))) +vuint32m1_t vredor(vuint32m1_t op0, vuint32m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m8_u32m1_m))) +vuint32m1_t vredor(vbool4_t op0, vuint32m1_t op1, vuint32m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32mf2_u32m1))) +vuint32m1_t vredor(vuint32m1_t op0, vuint32mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32mf2_u32m1_m))) +vuint32m1_t vredor(vbool64_t op0, vuint32m1_t op1, vuint32mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m1_u64m1))) +vuint64m1_t vredor(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m1_u64m1_m))) +vuint64m1_t vredor(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m2_u64m1))) +vuint64m1_t vredor(vuint64m1_t op0, vuint64m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m2_u64m1_m))) +vuint64m1_t vredor(vbool32_t op0, vuint64m1_t op1, vuint64m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m4_u64m1))) +vuint64m1_t vredor(vuint64m1_t op0, vuint64m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m4_u64m1_m))) +vuint64m1_t vredor(vbool16_t op0, vuint64m1_t op1, vuint64m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m8_u64m1))) +vuint64m1_t vredor(vuint64m1_t op0, vuint64m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m8_u64m1_m))) +vuint64m1_t vredor(vbool8_t op0, vuint64m1_t op1, vuint64m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m1_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m1_i8m1_m))) +vint8m1_t vredxor(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m2_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m2_i8m1_m))) +vint8m1_t vredxor(vbool4_t op0, vint8m1_t op1, vint8m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m4_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m4_i8m1_m))) +vint8m1_t vredxor(vbool2_t op0, vint8m1_t op1, vint8m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m8_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m8_i8m1_m))) +vint8m1_t vredxor(vbool1_t op0, vint8m1_t op1, vint8m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8mf2_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8mf2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8mf2_i8m1_m))) +vint8m1_t vredxor(vbool16_t op0, vint8m1_t op1, vint8mf2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8mf4_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8mf4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8mf4_i8m1_m))) +vint8m1_t vredxor(vbool32_t op0, vint8m1_t op1, vint8mf4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8mf8_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8mf8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8mf8_i8m1_m))) +vint8m1_t vredxor(vbool64_t op0, vint8m1_t op1, vint8mf8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m1_i16m1))) +vint16m1_t vredxor(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m1_i16m1_m))) +vint16m1_t vredxor(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m2_i16m1))) +vint16m1_t vredxor(vint16m1_t op0, vint16m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m2_i16m1_m))) +vint16m1_t vredxor(vbool8_t op0, vint16m1_t op1, vint16m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m4_i16m1))) +vint16m1_t vredxor(vint16m1_t op0, vint16m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m4_i16m1_m))) +vint16m1_t vredxor(vbool4_t op0, vint16m1_t op1, vint16m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m8_i16m1))) +vint16m1_t vredxor(vint16m1_t op0, vint16m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m8_i16m1_m))) +vint16m1_t vredxor(vbool2_t op0, vint16m1_t op1, vint16m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16mf2_i16m1))) +vint16m1_t vredxor(vint16m1_t op0, vint16mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16mf2_i16m1_m))) +vint16m1_t vredxor(vbool32_t op0, vint16m1_t op1, vint16mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16mf4_i16m1))) +vint16m1_t vredxor(vint16m1_t op0, vint16mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16mf4_i16m1_m))) +vint16m1_t vredxor(vbool64_t op0, vint16m1_t op1, vint16mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m1_i32m1))) +vint32m1_t vredxor(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m1_i32m1_m))) +vint32m1_t vredxor(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m2_i32m1))) +vint32m1_t vredxor(vint32m1_t op0, vint32m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m2_i32m1_m))) +vint32m1_t vredxor(vbool16_t op0, vint32m1_t op1, vint32m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m4_i32m1))) +vint32m1_t vredxor(vint32m1_t op0, vint32m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m4_i32m1_m))) +vint32m1_t vredxor(vbool8_t op0, vint32m1_t op1, vint32m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m8_i32m1))) +vint32m1_t vredxor(vint32m1_t op0, vint32m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m8_i32m1_m))) +vint32m1_t vredxor(vbool4_t op0, vint32m1_t op1, vint32m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32mf2_i32m1))) +vint32m1_t vredxor(vint32m1_t op0, vint32mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32mf2_i32m1_m))) +vint32m1_t vredxor(vbool64_t op0, vint32m1_t op1, vint32mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m1_i64m1))) +vint64m1_t vredxor(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m1_i64m1_m))) +vint64m1_t vredxor(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m2_i64m1))) +vint64m1_t vredxor(vint64m1_t op0, vint64m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m2_i64m1_m))) +vint64m1_t vredxor(vbool32_t op0, vint64m1_t op1, vint64m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m4_i64m1))) +vint64m1_t vredxor(vint64m1_t op0, vint64m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m4_i64m1_m))) +vint64m1_t vredxor(vbool16_t op0, vint64m1_t op1, vint64m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m8_i64m1))) +vint64m1_t vredxor(vint64m1_t op0, vint64m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m8_i64m1_m))) +vint64m1_t vredxor(vbool8_t op0, vint64m1_t op1, vint64m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m1_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m1_u8m1_m))) +vuint8m1_t vredxor(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m2_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m2_u8m1_m))) +vuint8m1_t vredxor(vbool4_t op0, vuint8m1_t op1, vuint8m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m4_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m4_u8m1_m))) +vuint8m1_t vredxor(vbool2_t op0, vuint8m1_t op1, vuint8m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m8_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m8_u8m1_m))) +vuint8m1_t vredxor(vbool1_t op0, vuint8m1_t op1, vuint8m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8mf2_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8mf2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8mf2_u8m1_m))) +vuint8m1_t vredxor(vbool16_t op0, vuint8m1_t op1, vuint8mf2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8mf4_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8mf4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8mf4_u8m1_m))) +vuint8m1_t vredxor(vbool32_t op0, vuint8m1_t op1, vuint8mf4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8mf8_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8mf8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8mf8_u8m1_m))) +vuint8m1_t vredxor(vbool64_t op0, vuint8m1_t op1, vuint8mf8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m1_u16m1))) +vuint16m1_t vredxor(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m1_u16m1_m))) +vuint16m1_t vredxor(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m2_u16m1))) +vuint16m1_t vredxor(vuint16m1_t op0, vuint16m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m2_u16m1_m))) +vuint16m1_t vredxor(vbool8_t op0, vuint16m1_t op1, vuint16m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m4_u16m1))) +vuint16m1_t vredxor(vuint16m1_t op0, vuint16m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m4_u16m1_m))) +vuint16m1_t vredxor(vbool4_t op0, vuint16m1_t op1, vuint16m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m8_u16m1))) +vuint16m1_t vredxor(vuint16m1_t op0, vuint16m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m8_u16m1_m))) +vuint16m1_t vredxor(vbool2_t op0, vuint16m1_t op1, vuint16m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16mf2_u16m1))) +vuint16m1_t vredxor(vuint16m1_t op0, vuint16mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16mf2_u16m1_m))) +vuint16m1_t vredxor(vbool32_t op0, vuint16m1_t op1, vuint16mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16mf4_u16m1))) +vuint16m1_t vredxor(vuint16m1_t op0, vuint16mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16mf4_u16m1_m))) +vuint16m1_t vredxor(vbool64_t op0, vuint16m1_t op1, vuint16mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m1_u32m1))) +vuint32m1_t vredxor(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m1_u32m1_m))) +vuint32m1_t vredxor(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m2_u32m1))) +vuint32m1_t vredxor(vuint32m1_t op0, vuint32m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m2_u32m1_m))) +vuint32m1_t vredxor(vbool16_t op0, vuint32m1_t op1, vuint32m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m4_u32m1))) +vuint32m1_t vredxor(vuint32m1_t op0, vuint32m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m4_u32m1_m))) +vuint32m1_t vredxor(vbool8_t op0, vuint32m1_t op1, vuint32m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m8_u32m1))) +vuint32m1_t vredxor(vuint32m1_t op0, vuint32m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m8_u32m1_m))) +vuint32m1_t vredxor(vbool4_t op0, vuint32m1_t op1, vuint32m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32mf2_u32m1))) +vuint32m1_t vredxor(vuint32m1_t op0, vuint32mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32mf2_u32m1_m))) +vuint32m1_t vredxor(vbool64_t op0, vuint32m1_t op1, vuint32mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m1_u64m1))) +vuint64m1_t vredxor(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m1_u64m1_m))) +vuint64m1_t vredxor(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m2_u64m1))) +vuint64m1_t vredxor(vuint64m1_t op0, vuint64m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m2_u64m1_m))) +vuint64m1_t vredxor(vbool32_t op0, vuint64m1_t op1, vuint64m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m4_u64m1))) +vuint64m1_t vredxor(vuint64m1_t op0, vuint64m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m4_u64m1_m))) +vuint64m1_t vredxor(vbool16_t op0, vuint64m1_t op1, vuint64m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m8_u64m1))) +vuint64m1_t vredxor(vuint64m1_t op0, vuint64m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m8_u64m1_m))) +vuint64m1_t vredxor(vbool8_t op0, vuint64m1_t op1, vuint64m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m1_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m1_i16m1_m))) +vint16m1_t vwredsum(vbool8_t op0, vint16m1_t op1, vint8m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m2_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m2_i16m1_m))) +vint16m1_t vwredsum(vbool4_t op0, vint16m1_t op1, vint8m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m4_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m4_i16m1_m))) +vint16m1_t vwredsum(vbool2_t op0, vint16m1_t op1, vint8m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m8_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m8_i16m1_m))) +vint16m1_t vwredsum(vbool1_t op0, vint16m1_t op1, vint8m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8mf2_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8mf2_i16m1_m))) +vint16m1_t vwredsum(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8mf4_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8mf4_i16m1_m))) +vint16m1_t vwredsum(vbool32_t op0, vint16m1_t op1, vint8mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8mf8_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8mf8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8mf8_i16m1_m))) +vint16m1_t vwredsum(vbool64_t op0, vint16m1_t op1, vint8mf8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m1_i32m1))) +vint32m1_t vwredsum(vint32m1_t op0, vint16m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m1_i32m1_m))) +vint32m1_t vwredsum(vbool16_t op0, vint32m1_t op1, vint16m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m2_i32m1))) +vint32m1_t vwredsum(vint32m1_t op0, vint16m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m2_i32m1_m))) +vint32m1_t vwredsum(vbool8_t op0, vint32m1_t op1, vint16m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m4_i32m1))) +vint32m1_t vwredsum(vint32m1_t op0, vint16m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m4_i32m1_m))) +vint32m1_t vwredsum(vbool4_t op0, vint32m1_t op1, vint16m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m8_i32m1))) +vint32m1_t vwredsum(vint32m1_t op0, vint16m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m8_i32m1_m))) +vint32m1_t vwredsum(vbool2_t op0, vint32m1_t op1, vint16m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16mf2_i32m1))) +vint32m1_t vwredsum(vint32m1_t op0, vint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16mf2_i32m1_m))) +vint32m1_t vwredsum(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16mf4_i32m1))) +vint32m1_t vwredsum(vint32m1_t op0, vint16mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16mf4_i32m1_m))) +vint32m1_t vwredsum(vbool64_t op0, vint32m1_t op1, vint16mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m1_i64m1))) +vint64m1_t vwredsum(vint64m1_t op0, vint32m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m1_i64m1_m))) +vint64m1_t vwredsum(vbool32_t op0, vint64m1_t op1, vint32m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m2_i64m1))) +vint64m1_t vwredsum(vint64m1_t op0, vint32m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m2_i64m1_m))) +vint64m1_t vwredsum(vbool16_t op0, vint64m1_t op1, vint32m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m4_i64m1))) +vint64m1_t vwredsum(vint64m1_t op0, vint32m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m4_i64m1_m))) +vint64m1_t vwredsum(vbool8_t op0, vint64m1_t op1, vint32m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m8_i64m1))) +vint64m1_t vwredsum(vint64m1_t op0, vint32m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m8_i64m1_m))) +vint64m1_t vwredsum(vbool4_t op0, vint64m1_t op1, vint32m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32mf2_i64m1))) +vint64m1_t vwredsum(vint64m1_t op0, vint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32mf2_i64m1_m))) +vint64m1_t vwredsum(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m1_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m1_u16m1_m))) +vuint16m1_t vwredsumu(vbool8_t op0, vuint16m1_t op1, vuint8m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m2_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m2_u16m1_m))) +vuint16m1_t vwredsumu(vbool4_t op0, vuint16m1_t op1, vuint8m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m4_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m4_u16m1_m))) +vuint16m1_t vwredsumu(vbool2_t op0, vuint16m1_t op1, vuint8m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m8_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m8_u16m1_m))) +vuint16m1_t vwredsumu(vbool1_t op0, vuint16m1_t op1, vuint8m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8mf2_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8mf2_u16m1_m))) +vuint16m1_t vwredsumu(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8mf4_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8mf4_u16m1_m))) +vuint16m1_t vwredsumu(vbool32_t op0, vuint16m1_t op1, vuint8mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8mf8_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8mf8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8mf8_u16m1_m))) +vuint16m1_t vwredsumu(vbool64_t op0, vuint16m1_t op1, vuint8mf8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m1_u32m1))) +vuint32m1_t vwredsumu(vuint32m1_t op0, vuint16m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m1_u32m1_m))) +vuint32m1_t vwredsumu(vbool16_t op0, vuint32m1_t op1, vuint16m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m2_u32m1))) +vuint32m1_t vwredsumu(vuint32m1_t op0, vuint16m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m2_u32m1_m))) +vuint32m1_t vwredsumu(vbool8_t op0, vuint32m1_t op1, vuint16m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m4_u32m1))) +vuint32m1_t vwredsumu(vuint32m1_t op0, vuint16m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m4_u32m1_m))) +vuint32m1_t vwredsumu(vbool4_t op0, vuint32m1_t op1, vuint16m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m8_u32m1))) +vuint32m1_t vwredsumu(vuint32m1_t op0, vuint16m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m8_u32m1_m))) +vuint32m1_t vwredsumu(vbool2_t op0, vuint32m1_t op1, vuint16m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16mf2_u32m1))) +vuint32m1_t vwredsumu(vuint32m1_t op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16mf2_u32m1_m))) +vuint32m1_t vwredsumu(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16mf4_u32m1))) +vuint32m1_t vwredsumu(vuint32m1_t op0, vuint16mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16mf4_u32m1_m))) +vuint32m1_t vwredsumu(vbool64_t op0, vuint32m1_t op1, vuint16mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m1_u64m1))) +vuint64m1_t vwredsumu(vuint64m1_t op0, vuint32m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m1_u64m1_m))) +vuint64m1_t vwredsumu(vbool32_t op0, vuint64m1_t op1, vuint32m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m2_u64m1))) +vuint64m1_t vwredsumu(vuint64m1_t op0, vuint32m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m2_u64m1_m))) +vuint64m1_t vwredsumu(vbool16_t op0, vuint64m1_t op1, vuint32m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m4_u64m1))) +vuint64m1_t vwredsumu(vuint64m1_t op0, vuint32m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m4_u64m1_m))) +vuint64m1_t vwredsumu(vbool8_t op0, vuint64m1_t op1, vuint32m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m8_u64m1))) +vuint64m1_t vwredsumu(vuint64m1_t op0, vuint32m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m8_u64m1_m))) +vuint64m1_t vwredsumu(vbool4_t op0, vuint64m1_t op1, vuint32m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32mf2_u64m1))) +vuint64m1_t vwredsumu(vuint64m1_t op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32mf2_u64m1_m))) +vuint64m1_t vwredsumu(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m1))) +void vsuxei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m1_m))) +void vsuxei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m2))) +void vsuxei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m2_m))) +void vsuxei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m4))) +void vsuxei8(uint8_t * op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m4_m))) +void vsuxei8(vbool2_t op0, uint8_t * op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m8))) +void vsuxei8(uint8_t * op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m8_m))) +void vsuxei8(vbool1_t op0, uint8_t * op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8mf2))) +void vsuxei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8mf2_m))) +void vsuxei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8mf4))) +void vsuxei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8mf4_m))) +void vsuxei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8mf8))) +void vsuxei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8mf8_m))) +void vsuxei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b8))) +vbool8_t vmnot(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b4))) +vbool4_t vmnot(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b2))) +vbool2_t vmnot(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b1))) +vbool1_t vmnot(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b16))) +vbool16_t vmnot(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b32))) +vbool32_t vmnot(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b64))) +vbool64_t vmnot(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8m1_m))) +vuint8m1_t viota(vbool8_t op0, vuint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8m2_m))) +vuint8m2_t viota(vbool4_t op0, vuint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8m4_m))) +vuint8m4_t viota(vbool2_t op0, vuint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8m8_m))) +vuint8m8_t viota(vbool1_t op0, vuint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8mf2_m))) +vuint8mf2_t viota(vbool16_t op0, vuint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8mf4_m))) +vuint8mf4_t viota(vbool32_t op0, vuint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8mf8_m))) +vuint8mf8_t viota(vbool64_t op0, vuint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u16m1_m))) +vuint16m1_t viota(vbool16_t op0, vuint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u16m2_m))) +vuint16m2_t viota(vbool8_t op0, vuint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u16m4_m))) +vuint16m4_t viota(vbool4_t op0, vuint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u16m8_m))) +vuint16m8_t viota(vbool2_t op0, vuint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u16mf2_m))) +vuint16mf2_t viota(vbool32_t op0, vuint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u16mf4_m))) +vuint16mf4_t viota(vbool64_t op0, vuint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u32m1_m))) +vuint32m1_t viota(vbool32_t op0, vuint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u32m2_m))) +vuint32m2_t viota(vbool16_t op0, vuint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u32m4_m))) +vuint32m4_t viota(vbool8_t op0, vuint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u32m8_m))) +vuint32m8_t viota(vbool4_t op0, vuint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u32mf2_m))) +vuint32mf2_t viota(vbool64_t op0, vuint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u64m1_m))) +vuint64m1_t viota(vbool64_t op0, vuint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u64m2_m))) +vuint64m2_t viota(vbool32_t op0, vuint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u64m4_m))) +vuint64m4_t viota(vbool16_t op0, vuint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u64m8_m))) +vuint64m8_t viota(vbool8_t op0, vuint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8m1_m))) +vint8m1_t vid(vbool8_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8m2_m))) +vint8m2_t vid(vbool4_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8m4_m))) +vint8m4_t vid(vbool2_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8m8_m))) +vint8m8_t vid(vbool1_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8mf2_m))) +vint8mf2_t vid(vbool16_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8mf4_m))) +vint8mf4_t vid(vbool32_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8mf8_m))) +vint8mf8_t vid(vbool64_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i16m1_m))) +vint16m1_t vid(vbool16_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i16m2_m))) +vint16m2_t vid(vbool8_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i16m4_m))) +vint16m4_t vid(vbool4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i16m8_m))) +vint16m8_t vid(vbool2_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i16mf2_m))) +vint16mf2_t vid(vbool32_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i16mf4_m))) +vint16mf4_t vid(vbool64_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i32m1_m))) +vint32m1_t vid(vbool32_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i32m2_m))) +vint32m2_t vid(vbool16_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i32m4_m))) +vint32m4_t vid(vbool8_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i32m8_m))) +vint32m8_t vid(vbool4_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i32mf2_m))) +vint32mf2_t vid(vbool64_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i64m1_m))) +vint64m1_t vid(vbool64_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i64m2_m))) +vint64m2_t vid(vbool32_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i64m4_m))) +vint64m4_t vid(vbool16_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i64m8_m))) +vint64m8_t vid(vbool8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8m1_m))) +vuint8m1_t vid(vbool8_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8m2_m))) +vuint8m2_t vid(vbool4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8m4_m))) +vuint8m4_t vid(vbool2_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8m8_m))) +vuint8m8_t vid(vbool1_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8mf2_m))) +vuint8mf2_t vid(vbool16_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8mf4_m))) +vuint8mf4_t vid(vbool32_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8mf8_m))) +vuint8mf8_t vid(vbool64_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u16m1_m))) +vuint16m1_t vid(vbool16_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u16m2_m))) +vuint16m2_t vid(vbool8_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u16m4_m))) +vuint16m4_t vid(vbool4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u16m8_m))) +vuint16m8_t vid(vbool2_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u16mf2_m))) +vuint16mf2_t vid(vbool32_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u16mf4_m))) +vuint16mf4_t vid(vbool64_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u32m1_m))) +vuint32m1_t vid(vbool32_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u32m2_m))) +vuint32m2_t vid(vbool16_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u32m4_m))) +vuint32m4_t vid(vbool8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u32m8_m))) +vuint32m8_t vid(vbool4_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u32mf2_m))) +vuint32mf2_t vid(vbool64_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u64m1_m))) +vuint64m1_t vid(vbool64_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u64m2_m))) +vuint64m2_t vid(vbool32_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u64m4_m))) +vuint64m4_t vid(vbool16_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u64m8_m))) +vuint64m8_t vid(vbool8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8m1_i8))) +int8_t vmv_x(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8m2_i8))) +int8_t vmv_x(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8m4_i8))) +int8_t vmv_x(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8m8_i8))) +int8_t vmv_x(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8mf2_i8))) +int8_t vmv_x(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8mf4_i8))) +int8_t vmv_x(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8mf8_i8))) +int8_t vmv_x(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i16m1_i16))) +int16_t vmv_x(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i16m2_i16))) +int16_t vmv_x(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i16m4_i16))) +int16_t vmv_x(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i16m8_i16))) +int16_t vmv_x(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i16mf2_i16))) +int16_t vmv_x(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i16mf4_i16))) +int16_t vmv_x(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i32m1_i32))) +int32_t vmv_x(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i32m2_i32))) +int32_t vmv_x(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i32m4_i32))) +int32_t vmv_x(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i32m8_i32))) +int32_t vmv_x(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i32mf2_i32))) +int32_t vmv_x(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i64m1_i64))) +int64_t vmv_x(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i64m2_i64))) +int64_t vmv_x(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i64m4_i64))) +int64_t vmv_x(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i64m8_i64))) +int64_t vmv_x(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8m1_u8))) +uint8_t vmv_x(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8m2_u8))) +uint8_t vmv_x(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8m4_u8))) +uint8_t vmv_x(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8m8_u8))) +uint8_t vmv_x(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8mf2_u8))) +uint8_t vmv_x(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8mf4_u8))) +uint8_t vmv_x(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8mf8_u8))) +uint8_t vmv_x(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u16m1_u16))) +uint16_t vmv_x(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u16m2_u16))) +uint16_t vmv_x(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u16m4_u16))) +uint16_t vmv_x(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u16m8_u16))) +uint16_t vmv_x(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u16mf2_u16))) +uint16_t vmv_x(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u16mf4_u16))) +uint16_t vmv_x(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u32m1_u32))) +uint32_t vmv_x(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u32m2_u32))) +uint32_t vmv_x(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u32m4_u32))) +uint32_t vmv_x(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u32m8_u32))) +uint32_t vmv_x(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u32mf2_u32))) +uint32_t vmv_x(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u64m1_u64))) +uint64_t vmv_x(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u64m2_u64))) +uint64_t vmv_x(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u64m4_u64))) +uint64_t vmv_x(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u64m8_u64))) +uint64_t vmv_x(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8m1))) +vint8m1_t vmv_s(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8m2))) +vint8m2_t vmv_s(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8m4))) +vint8m4_t vmv_s(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8m8))) +vint8m8_t vmv_s(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8mf2))) +vint8mf2_t vmv_s(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8mf4))) +vint8mf4_t vmv_s(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8mf8))) +vint8mf8_t vmv_s(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i16m1))) +vint16m1_t vmv_s(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i16m2))) +vint16m2_t vmv_s(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i16m4))) +vint16m4_t vmv_s(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i16m8))) +vint16m8_t vmv_s(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i16mf2))) +vint16mf2_t vmv_s(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i16mf4))) +vint16mf4_t vmv_s(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i32m1))) +vint32m1_t vmv_s(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i32m2))) +vint32m2_t vmv_s(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i32m4))) +vint32m4_t vmv_s(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i32m8))) +vint32m8_t vmv_s(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i32mf2))) +vint32mf2_t vmv_s(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i64m1))) +vint64m1_t vmv_s(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i64m2))) +vint64m2_t vmv_s(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i64m4))) +vint64m4_t vmv_s(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i64m8))) +vint64m8_t vmv_s(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8m1))) +vuint8m1_t vmv_s(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8m2))) +vuint8m2_t vmv_s(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8m4))) +vuint8m4_t vmv_s(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8m8))) +vuint8m8_t vmv_s(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8mf2))) +vuint8mf2_t vmv_s(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8mf4))) +vuint8mf4_t vmv_s(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8mf8))) +vuint8mf8_t vmv_s(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u16m1))) +vuint16m1_t vmv_s(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u16m2))) +vuint16m2_t vmv_s(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u16m4))) +vuint16m4_t vmv_s(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u16m8))) +vuint16m8_t vmv_s(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u16mf2))) +vuint16mf2_t vmv_s(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u16mf4))) +vuint16mf4_t vmv_s(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u32m1))) +vuint32m1_t vmv_s(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u32m2))) +vuint32m2_t vmv_s(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u32m4))) +vuint32m4_t vmv_s(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u32m8))) +vuint32m8_t vmv_s(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u32mf2))) +vuint32mf2_t vmv_s(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u64m1))) +vuint64m1_t vmv_s(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u64m2))) +vuint64m2_t vmv_s(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u64m4))) +vuint64m4_t vmv_s(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u64m8))) +vuint64m8_t vmv_s(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u16m1))) +vuint16m1_t vfmv_s(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u16m2))) +vuint16m2_t vfmv_s(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u16m4))) +vuint16m4_t vfmv_s(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u16m8))) +vuint16m8_t vfmv_s(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u16mf2))) +vuint16mf2_t vfmv_s(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u16mf4))) +vuint16mf4_t vfmv_s(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u32m1))) +vuint32m1_t vfmv_s(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u32m2))) +vuint32m2_t vfmv_s(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u32m4))) +vuint32m4_t vfmv_s(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u32m8))) +vuint32m8_t vfmv_s(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u32mf2))) +vuint32mf2_t vfmv_s(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u64m1))) +vuint64m1_t vfmv_s(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u64m2))) +vuint64m2_t vfmv_s(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u64m4))) +vuint64m4_t vfmv_s(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u64m8))) +vuint64m8_t vfmv_s(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m1))) +vint8m1_t vslideup(vint8m1_t op0, vint8m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m1_m))) +vint8m1_t vslideup(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m2))) +vint8m2_t vslideup(vint8m2_t op0, vint8m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m2_m))) +vint8m2_t vslideup(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m4))) +vint8m4_t vslideup(vint8m4_t op0, vint8m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m4_m))) +vint8m4_t vslideup(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m8))) +vint8m8_t vslideup(vint8m8_t op0, vint8m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m8_m))) +vint8m8_t vslideup(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8mf2))) +vint8mf2_t vslideup(vint8mf2_t op0, vint8mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8mf2_m))) +vint8mf2_t vslideup(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8mf4))) +vint8mf4_t vslideup(vint8mf4_t op0, vint8mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8mf4_m))) +vint8mf4_t vslideup(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8mf8))) +vint8mf8_t vslideup(vint8mf8_t op0, vint8mf8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8mf8_m))) +vint8mf8_t vslideup(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m1))) +vint16m1_t vslideup(vint16m1_t op0, vint16m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m1_m))) +vint16m1_t vslideup(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m2))) +vint16m2_t vslideup(vint16m2_t op0, vint16m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m2_m))) +vint16m2_t vslideup(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m4))) +vint16m4_t vslideup(vint16m4_t op0, vint16m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m4_m))) +vint16m4_t vslideup(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m8))) +vint16m8_t vslideup(vint16m8_t op0, vint16m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m8_m))) +vint16m8_t vslideup(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16mf2))) +vint16mf2_t vslideup(vint16mf2_t op0, vint16mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16mf2_m))) +vint16mf2_t vslideup(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16mf4))) +vint16mf4_t vslideup(vint16mf4_t op0, vint16mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16mf4_m))) +vint16mf4_t vslideup(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m1))) +vint32m1_t vslideup(vint32m1_t op0, vint32m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m1_m))) +vint32m1_t vslideup(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m2))) +vint32m2_t vslideup(vint32m2_t op0, vint32m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m2_m))) +vint32m2_t vslideup(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m4))) +vint32m4_t vslideup(vint32m4_t op0, vint32m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m4_m))) +vint32m4_t vslideup(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m8))) +vint32m8_t vslideup(vint32m8_t op0, vint32m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m8_m))) +vint32m8_t vslideup(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32mf2))) +vint32mf2_t vslideup(vint32mf2_t op0, vint32mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32mf2_m))) +vint32mf2_t vslideup(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m1))) +vint64m1_t vslideup(vint64m1_t op0, vint64m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m1_m))) +vint64m1_t vslideup(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m2))) +vint64m2_t vslideup(vint64m2_t op0, vint64m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m2_m))) +vint64m2_t vslideup(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m4))) +vint64m4_t vslideup(vint64m4_t op0, vint64m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m4_m))) +vint64m4_t vslideup(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m8))) +vint64m8_t vslideup(vint64m8_t op0, vint64m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m8_m))) +vint64m8_t vslideup(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m1))) +vuint8m1_t vslideup(vuint8m1_t op0, vuint8m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m1_m))) +vuint8m1_t vslideup(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m2))) +vuint8m2_t vslideup(vuint8m2_t op0, vuint8m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m2_m))) +vuint8m2_t vslideup(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m4))) +vuint8m4_t vslideup(vuint8m4_t op0, vuint8m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m4_m))) +vuint8m4_t vslideup(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m8))) +vuint8m8_t vslideup(vuint8m8_t op0, vuint8m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m8_m))) +vuint8m8_t vslideup(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8mf2))) +vuint8mf2_t vslideup(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8mf2_m))) +vuint8mf2_t vslideup(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8mf4))) +vuint8mf4_t vslideup(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8mf4_m))) +vuint8mf4_t vslideup(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8mf8))) +vuint8mf8_t vslideup(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8mf8_m))) +vuint8mf8_t vslideup(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m1))) +vuint16m1_t vslideup(vuint16m1_t op0, vuint16m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m1_m))) +vuint16m1_t vslideup(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m2))) +vuint16m2_t vslideup(vuint16m2_t op0, vuint16m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m2_m))) +vuint16m2_t vslideup(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m4))) +vuint16m4_t vslideup(vuint16m4_t op0, vuint16m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m4_m))) +vuint16m4_t vslideup(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m8))) +vuint16m8_t vslideup(vuint16m8_t op0, vuint16m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m8_m))) +vuint16m8_t vslideup(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16mf2))) +vuint16mf2_t vslideup(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16mf2_m))) +vuint16mf2_t vslideup(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16mf4))) +vuint16mf4_t vslideup(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16mf4_m))) +vuint16mf4_t vslideup(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m1))) +vuint32m1_t vslideup(vuint32m1_t op0, vuint32m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m1_m))) +vuint32m1_t vslideup(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m2))) +vuint32m2_t vslideup(vuint32m2_t op0, vuint32m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m2_m))) +vuint32m2_t vslideup(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m4))) +vuint32m4_t vslideup(vuint32m4_t op0, vuint32m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m4_m))) +vuint32m4_t vslideup(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m8))) +vuint32m8_t vslideup(vuint32m8_t op0, vuint32m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m8_m))) +vuint32m8_t vslideup(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32mf2))) +vuint32mf2_t vslideup(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32mf2_m))) +vuint32mf2_t vslideup(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m1))) +vuint64m1_t vslideup(vuint64m1_t op0, vuint64m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m1_m))) +vuint64m1_t vslideup(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m2))) +vuint64m2_t vslideup(vuint64m2_t op0, vuint64m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m2_m))) +vuint64m2_t vslideup(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m4))) +vuint64m4_t vslideup(vuint64m4_t op0, vuint64m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m4_m))) +vuint64m4_t vslideup(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m8))) +vuint64m8_t vslideup(vuint64m8_t op0, vuint64m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m8_m))) +vuint64m8_t vslideup(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m1))) +vint8m1_t vslidedown(vint8m1_t op0, vint8m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m1_m))) +vint8m1_t vslidedown(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m2))) +vint8m2_t vslidedown(vint8m2_t op0, vint8m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m2_m))) +vint8m2_t vslidedown(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m4))) +vint8m4_t vslidedown(vint8m4_t op0, vint8m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m4_m))) +vint8m4_t vslidedown(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m8))) +vint8m8_t vslidedown(vint8m8_t op0, vint8m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m8_m))) +vint8m8_t vslidedown(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8mf2))) +vint8mf2_t vslidedown(vint8mf2_t op0, vint8mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8mf2_m))) +vint8mf2_t vslidedown(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8mf4))) +vint8mf4_t vslidedown(vint8mf4_t op0, vint8mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8mf4_m))) +vint8mf4_t vslidedown(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8mf8))) +vint8mf8_t vslidedown(vint8mf8_t op0, vint8mf8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8mf8_m))) +vint8mf8_t vslidedown(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m1))) +vint16m1_t vslidedown(vint16m1_t op0, vint16m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m1_m))) +vint16m1_t vslidedown(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m2))) +vint16m2_t vslidedown(vint16m2_t op0, vint16m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m2_m))) +vint16m2_t vslidedown(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m4))) +vint16m4_t vslidedown(vint16m4_t op0, vint16m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m4_m))) +vint16m4_t vslidedown(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m8))) +vint16m8_t vslidedown(vint16m8_t op0, vint16m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m8_m))) +vint16m8_t vslidedown(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16mf2))) +vint16mf2_t vslidedown(vint16mf2_t op0, vint16mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16mf2_m))) +vint16mf2_t vslidedown(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16mf4))) +vint16mf4_t vslidedown(vint16mf4_t op0, vint16mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16mf4_m))) +vint16mf4_t vslidedown(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m1))) +vint32m1_t vslidedown(vint32m1_t op0, vint32m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m1_m))) +vint32m1_t vslidedown(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m2))) +vint32m2_t vslidedown(vint32m2_t op0, vint32m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m2_m))) +vint32m2_t vslidedown(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m4))) +vint32m4_t vslidedown(vint32m4_t op0, vint32m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m4_m))) +vint32m4_t vslidedown(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m8))) +vint32m8_t vslidedown(vint32m8_t op0, vint32m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m8_m))) +vint32m8_t vslidedown(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32mf2))) +vint32mf2_t vslidedown(vint32mf2_t op0, vint32mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32mf2_m))) +vint32mf2_t vslidedown(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m1))) +vint64m1_t vslidedown(vint64m1_t op0, vint64m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m1_m))) +vint64m1_t vslidedown(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m2))) +vint64m2_t vslidedown(vint64m2_t op0, vint64m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m2_m))) +vint64m2_t vslidedown(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m4))) +vint64m4_t vslidedown(vint64m4_t op0, vint64m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m4_m))) +vint64m4_t vslidedown(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m8))) +vint64m8_t vslidedown(vint64m8_t op0, vint64m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m8_m))) +vint64m8_t vslidedown(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m1))) +vuint8m1_t vslidedown(vuint8m1_t op0, vuint8m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m1_m))) +vuint8m1_t vslidedown(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m2))) +vuint8m2_t vslidedown(vuint8m2_t op0, vuint8m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m2_m))) +vuint8m2_t vslidedown(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m4))) +vuint8m4_t vslidedown(vuint8m4_t op0, vuint8m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m4_m))) +vuint8m4_t vslidedown(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m8))) +vuint8m8_t vslidedown(vuint8m8_t op0, vuint8m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m8_m))) +vuint8m8_t vslidedown(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8mf2))) +vuint8mf2_t vslidedown(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8mf2_m))) +vuint8mf2_t vslidedown(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8mf4))) +vuint8mf4_t vslidedown(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8mf4_m))) +vuint8mf4_t vslidedown(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8mf8))) +vuint8mf8_t vslidedown(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8mf8_m))) +vuint8mf8_t vslidedown(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m1))) +vuint16m1_t vslidedown(vuint16m1_t op0, vuint16m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m1_m))) +vuint16m1_t vslidedown(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m2))) +vuint16m2_t vslidedown(vuint16m2_t op0, vuint16m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m2_m))) +vuint16m2_t vslidedown(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m4))) +vuint16m4_t vslidedown(vuint16m4_t op0, vuint16m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m4_m))) +vuint16m4_t vslidedown(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m8))) +vuint16m8_t vslidedown(vuint16m8_t op0, vuint16m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m8_m))) +vuint16m8_t vslidedown(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16mf2))) +vuint16mf2_t vslidedown(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16mf2_m))) +vuint16mf2_t vslidedown(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16mf4))) +vuint16mf4_t vslidedown(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16mf4_m))) +vuint16mf4_t vslidedown(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m1))) +vuint32m1_t vslidedown(vuint32m1_t op0, vuint32m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m1_m))) +vuint32m1_t vslidedown(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m2))) +vuint32m2_t vslidedown(vuint32m2_t op0, vuint32m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m2_m))) +vuint32m2_t vslidedown(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m4))) +vuint32m4_t vslidedown(vuint32m4_t op0, vuint32m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m4_m))) +vuint32m4_t vslidedown(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m8))) +vuint32m8_t vslidedown(vuint32m8_t op0, vuint32m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m8_m))) +vuint32m8_t vslidedown(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32mf2))) +vuint32mf2_t vslidedown(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32mf2_m))) +vuint32mf2_t vslidedown(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m1))) +vuint64m1_t vslidedown(vuint64m1_t op0, vuint64m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m1_m))) +vuint64m1_t vslidedown(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m2))) +vuint64m2_t vslidedown(vuint64m2_t op0, vuint64m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m2_m))) +vuint64m2_t vslidedown(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m4))) +vuint64m4_t vslidedown(vuint64m4_t op0, vuint64m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m4_m))) +vuint64m4_t vslidedown(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m8))) +vuint64m8_t vslidedown(vuint64m8_t op0, vuint64m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m8_m))) +vuint64m8_t vslidedown(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m1))) +vint8m1_t vslide1up(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m1_m))) +vint8m1_t vslide1up(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m2))) +vint8m2_t vslide1up(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m2_m))) +vint8m2_t vslide1up(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m4))) +vint8m4_t vslide1up(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m4_m))) +vint8m4_t vslide1up(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m8))) +vint8m8_t vslide1up(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m8_m))) +vint8m8_t vslide1up(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8mf2))) +vint8mf2_t vslide1up(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8mf2_m))) +vint8mf2_t vslide1up(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8mf4))) +vint8mf4_t vslide1up(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8mf4_m))) +vint8mf4_t vslide1up(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8mf8))) +vint8mf8_t vslide1up(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8mf8_m))) +vint8mf8_t vslide1up(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m1))) +vint16m1_t vslide1up(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m1_m))) +vint16m1_t vslide1up(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m2))) +vint16m2_t vslide1up(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m2_m))) +vint16m2_t vslide1up(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m4))) +vint16m4_t vslide1up(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m4_m))) +vint16m4_t vslide1up(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m8))) +vint16m8_t vslide1up(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m8_m))) +vint16m8_t vslide1up(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16mf2))) +vint16mf2_t vslide1up(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16mf2_m))) +vint16mf2_t vslide1up(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16mf4))) +vint16mf4_t vslide1up(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16mf4_m))) +vint16mf4_t vslide1up(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m1))) +vint32m1_t vslide1up(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m1_m))) +vint32m1_t vslide1up(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m2))) +vint32m2_t vslide1up(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m2_m))) +vint32m2_t vslide1up(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m4))) +vint32m4_t vslide1up(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m4_m))) +vint32m4_t vslide1up(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m8))) +vint32m8_t vslide1up(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m8_m))) +vint32m8_t vslide1up(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32mf2))) +vint32mf2_t vslide1up(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32mf2_m))) +vint32mf2_t vslide1up(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m1))) +vint64m1_t vslide1up(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m1_m))) +vint64m1_t vslide1up(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m2))) +vint64m2_t vslide1up(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m2_m))) +vint64m2_t vslide1up(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m4))) +vint64m4_t vslide1up(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m4_m))) +vint64m4_t vslide1up(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m8))) +vint64m8_t vslide1up(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m8_m))) +vint64m8_t vslide1up(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m1))) +vuint8m1_t vslide1up(vuint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m1_m))) +vuint8m1_t vslide1up(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m2))) +vuint8m2_t vslide1up(vuint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m2_m))) +vuint8m2_t vslide1up(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m4))) +vuint8m4_t vslide1up(vuint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m4_m))) +vuint8m4_t vslide1up(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m8))) +vuint8m8_t vslide1up(vuint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m8_m))) +vuint8m8_t vslide1up(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8mf2))) +vuint8mf2_t vslide1up(vuint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8mf2_m))) +vuint8mf2_t vslide1up(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8mf4))) +vuint8mf4_t vslide1up(vuint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8mf4_m))) +vuint8mf4_t vslide1up(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8mf8))) +vuint8mf8_t vslide1up(vuint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8mf8_m))) +vuint8mf8_t vslide1up(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m1))) +vuint16m1_t vslide1up(vuint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m1_m))) +vuint16m1_t vslide1up(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m2))) +vuint16m2_t vslide1up(vuint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m2_m))) +vuint16m2_t vslide1up(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m4))) +vuint16m4_t vslide1up(vuint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m4_m))) +vuint16m4_t vslide1up(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m8))) +vuint16m8_t vslide1up(vuint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m8_m))) +vuint16m8_t vslide1up(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16mf2))) +vuint16mf2_t vslide1up(vuint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16mf2_m))) +vuint16mf2_t vslide1up(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16mf4))) +vuint16mf4_t vslide1up(vuint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16mf4_m))) +vuint16mf4_t vslide1up(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m1))) +vuint32m1_t vslide1up(vuint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m1_m))) +vuint32m1_t vslide1up(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m2))) +vuint32m2_t vslide1up(vuint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m2_m))) +vuint32m2_t vslide1up(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m4))) +vuint32m4_t vslide1up(vuint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m4_m))) +vuint32m4_t vslide1up(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m8))) +vuint32m8_t vslide1up(vuint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m8_m))) +vuint32m8_t vslide1up(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32mf2))) +vuint32mf2_t vslide1up(vuint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32mf2_m))) +vuint32mf2_t vslide1up(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m1))) +vuint64m1_t vslide1up(vuint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m1_m))) +vuint64m1_t vslide1up(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m2))) +vuint64m2_t vslide1up(vuint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m2_m))) +vuint64m2_t vslide1up(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m4))) +vuint64m4_t vslide1up(vuint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m4_m))) +vuint64m4_t vslide1up(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m8))) +vuint64m8_t vslide1up(vuint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m8_m))) +vuint64m8_t vslide1up(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m1))) +vint8m1_t vslide1down(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m1_m))) +vint8m1_t vslide1down(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m2))) +vint8m2_t vslide1down(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m2_m))) +vint8m2_t vslide1down(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m4))) +vint8m4_t vslide1down(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m4_m))) +vint8m4_t vslide1down(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m8))) +vint8m8_t vslide1down(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m8_m))) +vint8m8_t vslide1down(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8mf2))) +vint8mf2_t vslide1down(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8mf2_m))) +vint8mf2_t vslide1down(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8mf4))) +vint8mf4_t vslide1down(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8mf4_m))) +vint8mf4_t vslide1down(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8mf8))) +vint8mf8_t vslide1down(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8mf8_m))) +vint8mf8_t vslide1down(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m1))) +vint16m1_t vslide1down(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m1_m))) +vint16m1_t vslide1down(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m2))) +vint16m2_t vslide1down(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m2_m))) +vint16m2_t vslide1down(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m4))) +vint16m4_t vslide1down(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m4_m))) +vint16m4_t vslide1down(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m8))) +vint16m8_t vslide1down(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m8_m))) +vint16m8_t vslide1down(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16mf2))) +vint16mf2_t vslide1down(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16mf2_m))) +vint16mf2_t vslide1down(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16mf4))) +vint16mf4_t vslide1down(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16mf4_m))) +vint16mf4_t vslide1down(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m1))) +vint32m1_t vslide1down(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m1_m))) +vint32m1_t vslide1down(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m2))) +vint32m2_t vslide1down(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m2_m))) +vint32m2_t vslide1down(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m4))) +vint32m4_t vslide1down(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m4_m))) +vint32m4_t vslide1down(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m8))) +vint32m8_t vslide1down(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m8_m))) +vint32m8_t vslide1down(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32mf2))) +vint32mf2_t vslide1down(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32mf2_m))) +vint32mf2_t vslide1down(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m1))) +vint64m1_t vslide1down(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m1_m))) +vint64m1_t vslide1down(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m2))) +vint64m2_t vslide1down(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m2_m))) +vint64m2_t vslide1down(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m4))) +vint64m4_t vslide1down(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m4_m))) +vint64m4_t vslide1down(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m8))) +vint64m8_t vslide1down(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m8_m))) +vint64m8_t vslide1down(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m1))) +vuint8m1_t vslide1down(vuint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m1_m))) +vuint8m1_t vslide1down(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m2))) +vuint8m2_t vslide1down(vuint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m2_m))) +vuint8m2_t vslide1down(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m4))) +vuint8m4_t vslide1down(vuint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m4_m))) +vuint8m4_t vslide1down(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m8))) +vuint8m8_t vslide1down(vuint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m8_m))) +vuint8m8_t vslide1down(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8mf2))) +vuint8mf2_t vslide1down(vuint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8mf2_m))) +vuint8mf2_t vslide1down(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8mf4))) +vuint8mf4_t vslide1down(vuint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8mf4_m))) +vuint8mf4_t vslide1down(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8mf8))) +vuint8mf8_t vslide1down(vuint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8mf8_m))) +vuint8mf8_t vslide1down(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m1))) +vuint16m1_t vslide1down(vuint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m1_m))) +vuint16m1_t vslide1down(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m2))) +vuint16m2_t vslide1down(vuint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m2_m))) +vuint16m2_t vslide1down(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m4))) +vuint16m4_t vslide1down(vuint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m4_m))) +vuint16m4_t vslide1down(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m8))) +vuint16m8_t vslide1down(vuint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m8_m))) +vuint16m8_t vslide1down(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16mf2))) +vuint16mf2_t vslide1down(vuint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16mf2_m))) +vuint16mf2_t vslide1down(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16mf4))) +vuint16mf4_t vslide1down(vuint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16mf4_m))) +vuint16mf4_t vslide1down(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m1))) +vuint32m1_t vslide1down(vuint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m1_m))) +vuint32m1_t vslide1down(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m2))) +vuint32m2_t vslide1down(vuint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m2_m))) +vuint32m2_t vslide1down(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m4))) +vuint32m4_t vslide1down(vuint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m4_m))) +vuint32m4_t vslide1down(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m8))) +vuint32m8_t vslide1down(vuint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m8_m))) +vuint32m8_t vslide1down(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32mf2))) +vuint32mf2_t vslide1down(vuint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32mf2_m))) +vuint32mf2_t vslide1down(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m1))) +vuint64m1_t vslide1down(vuint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m1_m))) +vuint64m1_t vslide1down(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m2))) +vuint64m2_t vslide1down(vuint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m2_m))) +vuint64m2_t vslide1down(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m4))) +vuint64m4_t vslide1down(vuint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m4_m))) +vuint64m4_t vslide1down(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m8))) +vuint64m8_t vslide1down(vuint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m8_m))) +vuint64m8_t vslide1down(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m1))) +vint8m1_t vrgather(vint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m1_m))) +vint8m1_t vrgather(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m2))) +vint8m2_t vrgather(vint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m2_m))) +vint8m2_t vrgather(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m4))) +vint8m4_t vrgather(vint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m4_m))) +vint8m4_t vrgather(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m8))) +vint8m8_t vrgather(vint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m8_m))) +vint8m8_t vrgather(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8mf2))) +vint8mf2_t vrgather(vint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8mf2_m))) +vint8mf2_t vrgather(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8mf4))) +vint8mf4_t vrgather(vint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8mf4_m))) +vint8mf4_t vrgather(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8mf8))) +vint8mf8_t vrgather(vint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8mf8_m))) +vint8mf8_t vrgather(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m1))) +vint16m1_t vrgather(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m1_m))) +vint16m1_t vrgather(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m2))) +vint16m2_t vrgather(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m2_m))) +vint16m2_t vrgather(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m4))) +vint16m4_t vrgather(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m4_m))) +vint16m4_t vrgather(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m8))) +vint16m8_t vrgather(vint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m8_m))) +vint16m8_t vrgather(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16mf2))) +vint16mf2_t vrgather(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16mf2_m))) +vint16mf2_t vrgather(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16mf4))) +vint16mf4_t vrgather(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16mf4_m))) +vint16mf4_t vrgather(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m1))) +vint32m1_t vrgather(vint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m1_m))) +vint32m1_t vrgather(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m2))) +vint32m2_t vrgather(vint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m2_m))) +vint32m2_t vrgather(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m4))) +vint32m4_t vrgather(vint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m4_m))) +vint32m4_t vrgather(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m8))) +vint32m8_t vrgather(vint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m8_m))) +vint32m8_t vrgather(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32mf2))) +vint32mf2_t vrgather(vint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32mf2_m))) +vint32mf2_t vrgather(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m1))) +vint64m1_t vrgather(vint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m1_m))) +vint64m1_t vrgather(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m2))) +vint64m2_t vrgather(vint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m2_m))) +vint64m2_t vrgather(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m4))) +vint64m4_t vrgather(vint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m4_m))) +vint64m4_t vrgather(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m8))) +vint64m8_t vrgather(vint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m8_m))) +vint64m8_t vrgather(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m1))) +vint8m1_t vrgather(vint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m1_m))) +vint8m1_t vrgather(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m2))) +vint8m2_t vrgather(vint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m2_m))) +vint8m2_t vrgather(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m4))) +vint8m4_t vrgather(vint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m4_m))) +vint8m4_t vrgather(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m8))) +vint8m8_t vrgather(vint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m8_m))) +vint8m8_t vrgather(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8mf2))) +vint8mf2_t vrgather(vint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8mf2_m))) +vint8mf2_t vrgather(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8mf4))) +vint8mf4_t vrgather(vint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8mf4_m))) +vint8mf4_t vrgather(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8mf8))) +vint8mf8_t vrgather(vint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8mf8_m))) +vint8mf8_t vrgather(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m1))) +vint16m1_t vrgather(vint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m1_m))) +vint16m1_t vrgather(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m2))) +vint16m2_t vrgather(vint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m2_m))) +vint16m2_t vrgather(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m4))) +vint16m4_t vrgather(vint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m4_m))) +vint16m4_t vrgather(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m8))) +vint16m8_t vrgather(vint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m8_m))) +vint16m8_t vrgather(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16mf2))) +vint16mf2_t vrgather(vint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16mf2_m))) +vint16mf2_t vrgather(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16mf4))) +vint16mf4_t vrgather(vint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16mf4_m))) +vint16mf4_t vrgather(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m1))) +vint32m1_t vrgather(vint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m1_m))) +vint32m1_t vrgather(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m2))) +vint32m2_t vrgather(vint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m2_m))) +vint32m2_t vrgather(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m4))) +vint32m4_t vrgather(vint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m4_m))) +vint32m4_t vrgather(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m8))) +vint32m8_t vrgather(vint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m8_m))) +vint32m8_t vrgather(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32mf2))) +vint32mf2_t vrgather(vint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32mf2_m))) +vint32mf2_t vrgather(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m1))) +vint64m1_t vrgather(vint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m1_m))) +vint64m1_t vrgather(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m2))) +vint64m2_t vrgather(vint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m2_m))) +vint64m2_t vrgather(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m4))) +vint64m4_t vrgather(vint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m4_m))) +vint64m4_t vrgather(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m8))) +vint64m8_t vrgather(vint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m8_m))) +vint64m8_t vrgather(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8m1))) +vint8m1_t vrgatherei16(vint8m1_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8m1_m))) +vint8m1_t vrgatherei16(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8m2))) +vint8m2_t vrgatherei16(vint8m2_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8m2_m))) +vint8m2_t vrgatherei16(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8m4))) +vint8m4_t vrgatherei16(vint8m4_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8m4_m))) +vint8m4_t vrgatherei16(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8mf2))) +vint8mf2_t vrgatherei16(vint8mf2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8mf2_m))) +vint8mf2_t vrgatherei16(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8mf4))) +vint8mf4_t vrgatherei16(vint8mf4_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8mf4_m))) +vint8mf4_t vrgatherei16(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8mf8))) +vint8mf8_t vrgatherei16(vint8mf8_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8mf8_m))) +vint8mf8_t vrgatherei16(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m1))) +vint16m1_t vrgatherei16(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m1_m))) +vint16m1_t vrgatherei16(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m2))) +vint16m2_t vrgatherei16(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m2_m))) +vint16m2_t vrgatherei16(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m4))) +vint16m4_t vrgatherei16(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m4_m))) +vint16m4_t vrgatherei16(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m8))) +vint16m8_t vrgatherei16(vint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m8_m))) +vint16m8_t vrgatherei16(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16mf2))) +vint16mf2_t vrgatherei16(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16mf2_m))) +vint16mf2_t vrgatherei16(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16mf4))) +vint16mf4_t vrgatherei16(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16mf4_m))) +vint16mf4_t vrgatherei16(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m1))) +vint32m1_t vrgatherei16(vint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m1_m))) +vint32m1_t vrgatherei16(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m2))) +vint32m2_t vrgatherei16(vint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m2_m))) +vint32m2_t vrgatherei16(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m4))) +vint32m4_t vrgatherei16(vint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m4_m))) +vint32m4_t vrgatherei16(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m8))) +vint32m8_t vrgatherei16(vint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m8_m))) +vint32m8_t vrgatherei16(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32mf2))) +vint32mf2_t vrgatherei16(vint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32mf2_m))) +vint32mf2_t vrgatherei16(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m1))) +vint64m1_t vrgatherei16(vint64m1_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m1_m))) +vint64m1_t vrgatherei16(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m2))) +vint64m2_t vrgatherei16(vint64m2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m2_m))) +vint64m2_t vrgatherei16(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m4))) +vint64m4_t vrgatherei16(vint64m4_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m4_m))) +vint64m4_t vrgatherei16(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m8))) +vint64m8_t vrgatherei16(vint64m8_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m8_m))) +vint64m8_t vrgatherei16(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m1))) +vuint8m1_t vrgather(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m1_m))) +vuint8m1_t vrgather(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m2))) +vuint8m2_t vrgather(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m2_m))) +vuint8m2_t vrgather(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m4))) +vuint8m4_t vrgather(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m4_m))) +vuint8m4_t vrgather(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m8))) +vuint8m8_t vrgather(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m8_m))) +vuint8m8_t vrgather(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8mf2))) +vuint8mf2_t vrgather(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8mf2_m))) +vuint8mf2_t vrgather(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8mf4))) +vuint8mf4_t vrgather(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8mf4_m))) +vuint8mf4_t vrgather(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8mf8))) +vuint8mf8_t vrgather(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8mf8_m))) +vuint8mf8_t vrgather(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m1))) +vuint16m1_t vrgather(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m1_m))) +vuint16m1_t vrgather(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m2))) +vuint16m2_t vrgather(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m2_m))) +vuint16m2_t vrgather(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m4))) +vuint16m4_t vrgather(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m4_m))) +vuint16m4_t vrgather(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m8))) +vuint16m8_t vrgather(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m8_m))) +vuint16m8_t vrgather(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16mf2))) +vuint16mf2_t vrgather(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16mf2_m))) +vuint16mf2_t vrgather(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16mf4))) +vuint16mf4_t vrgather(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16mf4_m))) +vuint16mf4_t vrgather(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m1))) +vuint32m1_t vrgather(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m1_m))) +vuint32m1_t vrgather(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m2))) +vuint32m2_t vrgather(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m2_m))) +vuint32m2_t vrgather(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m4))) +vuint32m4_t vrgather(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m4_m))) +vuint32m4_t vrgather(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m8))) +vuint32m8_t vrgather(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m8_m))) +vuint32m8_t vrgather(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32mf2))) +vuint32mf2_t vrgather(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32mf2_m))) +vuint32mf2_t vrgather(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m1))) +vuint64m1_t vrgather(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m1_m))) +vuint64m1_t vrgather(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m2))) +vuint64m2_t vrgather(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m2_m))) +vuint64m2_t vrgather(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m4))) +vuint64m4_t vrgather(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m4_m))) +vuint64m4_t vrgather(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m8))) +vuint64m8_t vrgather(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m8_m))) +vuint64m8_t vrgather(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m1))) +vuint8m1_t vrgather(vuint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m1_m))) +vuint8m1_t vrgather(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m2))) +vuint8m2_t vrgather(vuint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m2_m))) +vuint8m2_t vrgather(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m4))) +vuint8m4_t vrgather(vuint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m4_m))) +vuint8m4_t vrgather(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m8))) +vuint8m8_t vrgather(vuint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m8_m))) +vuint8m8_t vrgather(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8mf2))) +vuint8mf2_t vrgather(vuint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8mf2_m))) +vuint8mf2_t vrgather(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8mf4))) +vuint8mf4_t vrgather(vuint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8mf4_m))) +vuint8mf4_t vrgather(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8mf8))) +vuint8mf8_t vrgather(vuint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8mf8_m))) +vuint8mf8_t vrgather(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m1))) +vuint16m1_t vrgather(vuint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m1_m))) +vuint16m1_t vrgather(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m2))) +vuint16m2_t vrgather(vuint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m2_m))) +vuint16m2_t vrgather(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m4))) +vuint16m4_t vrgather(vuint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m4_m))) +vuint16m4_t vrgather(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m8))) +vuint16m8_t vrgather(vuint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m8_m))) +vuint16m8_t vrgather(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16mf2))) +vuint16mf2_t vrgather(vuint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16mf2_m))) +vuint16mf2_t vrgather(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16mf4))) +vuint16mf4_t vrgather(vuint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16mf4_m))) +vuint16mf4_t vrgather(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m1))) +vuint32m1_t vrgather(vuint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m1_m))) +vuint32m1_t vrgather(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m2))) +vuint32m2_t vrgather(vuint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m2_m))) +vuint32m2_t vrgather(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m4))) +vuint32m4_t vrgather(vuint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m4_m))) +vuint32m4_t vrgather(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m8))) +vuint32m8_t vrgather(vuint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m8_m))) +vuint32m8_t vrgather(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32mf2))) +vuint32mf2_t vrgather(vuint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32mf2_m))) +vuint32mf2_t vrgather(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m1))) +vuint64m1_t vrgather(vuint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m1_m))) +vuint64m1_t vrgather(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m2))) +vuint64m2_t vrgather(vuint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m2_m))) +vuint64m2_t vrgather(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m4))) +vuint64m4_t vrgather(vuint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m4_m))) +vuint64m4_t vrgather(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m8))) +vuint64m8_t vrgather(vuint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m8_m))) +vuint64m8_t vrgather(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8m1))) +vuint8m1_t vrgatherei16(vuint8m1_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8m1_m))) +vuint8m1_t vrgatherei16(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8m2))) +vuint8m2_t vrgatherei16(vuint8m2_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8m2_m))) +vuint8m2_t vrgatherei16(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8m4))) +vuint8m4_t vrgatherei16(vuint8m4_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8m4_m))) +vuint8m4_t vrgatherei16(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8mf2))) +vuint8mf2_t vrgatherei16(vuint8mf2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8mf2_m))) +vuint8mf2_t vrgatherei16(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8mf4))) +vuint8mf4_t vrgatherei16(vuint8mf4_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8mf4_m))) +vuint8mf4_t vrgatherei16(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8mf8))) +vuint8mf8_t vrgatherei16(vuint8mf8_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8mf8_m))) +vuint8mf8_t vrgatherei16(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m1))) +vuint16m1_t vrgatherei16(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m1_m))) +vuint16m1_t vrgatherei16(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m2))) +vuint16m2_t vrgatherei16(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m2_m))) +vuint16m2_t vrgatherei16(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m4))) +vuint16m4_t vrgatherei16(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m4_m))) +vuint16m4_t vrgatherei16(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m8))) +vuint16m8_t vrgatherei16(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m8_m))) +vuint16m8_t vrgatherei16(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16mf2))) +vuint16mf2_t vrgatherei16(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16mf2_m))) +vuint16mf2_t vrgatherei16(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16mf4))) +vuint16mf4_t vrgatherei16(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16mf4_m))) +vuint16mf4_t vrgatherei16(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m1))) +vuint32m1_t vrgatherei16(vuint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m1_m))) +vuint32m1_t vrgatherei16(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m2))) +vuint32m2_t vrgatherei16(vuint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m2_m))) +vuint32m2_t vrgatherei16(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m4))) +vuint32m4_t vrgatherei16(vuint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m4_m))) +vuint32m4_t vrgatherei16(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m8))) +vuint32m8_t vrgatherei16(vuint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m8_m))) +vuint32m8_t vrgatherei16(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32mf2))) +vuint32mf2_t vrgatherei16(vuint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32mf2_m))) +vuint32mf2_t vrgatherei16(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m1))) +vuint64m1_t vrgatherei16(vuint64m1_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m1_m))) +vuint64m1_t vrgatherei16(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m2))) +vuint64m2_t vrgatherei16(vuint64m2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m2_m))) +vuint64m2_t vrgatherei16(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m4))) +vuint64m4_t vrgatherei16(vuint64m4_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m4_m))) +vuint64m4_t vrgatherei16(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m8))) +vuint64m8_t vrgatherei16(vuint64m8_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m8_m))) +vuint64m8_t vrgatherei16(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8m1))) +vint8m1_t vcompress(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8m2))) +vint8m2_t vcompress(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8m4))) +vint8m4_t vcompress(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8m8))) +vint8m8_t vcompress(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8mf2))) +vint8mf2_t vcompress(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8mf4))) +vint8mf4_t vcompress(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8mf8))) +vint8mf8_t vcompress(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i16m1))) +vint16m1_t vcompress(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i16m2))) +vint16m2_t vcompress(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i16m4))) +vint16m4_t vcompress(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i16m8))) +vint16m8_t vcompress(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i16mf2))) +vint16mf2_t vcompress(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i16mf4))) +vint16mf4_t vcompress(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i32m1))) +vint32m1_t vcompress(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i32m2))) +vint32m2_t vcompress(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i32m4))) +vint32m4_t vcompress(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i32m8))) +vint32m8_t vcompress(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i32mf2))) +vint32mf2_t vcompress(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i64m1))) +vint64m1_t vcompress(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i64m2))) +vint64m2_t vcompress(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i64m4))) +vint64m4_t vcompress(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i64m8))) +vint64m8_t vcompress(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8m1))) +vuint8m1_t vcompress(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8m2))) +vuint8m2_t vcompress(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8m4))) +vuint8m4_t vcompress(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8m8))) +vuint8m8_t vcompress(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8mf2))) +vuint8mf2_t vcompress(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8mf4))) +vuint8mf4_t vcompress(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8mf8))) +vuint8mf8_t vcompress(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u16m1))) +vuint16m1_t vcompress(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u16m2))) +vuint16m2_t vcompress(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u16m4))) +vuint16m4_t vcompress(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u16m8))) +vuint16m8_t vcompress(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u16mf2))) +vuint16mf2_t vcompress(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u16mf4))) +vuint16mf4_t vcompress(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u32m1))) +vuint32m1_t vcompress(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u32m2))) +vuint32m2_t vcompress(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u32m4))) +vuint32m4_t vcompress(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u32m8))) +vuint32m8_t vcompress(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u32mf2))) +vuint32mf2_t vcompress(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u64m1))) +vuint64m1_t vcompress(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u64m2))) +vuint64m2_t vcompress(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u64m4))) +vuint64m4_t vcompress(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u64m8))) +vuint64m8_t vcompress(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i8m2_i8m1))) +vint8m1_t vget_i8m1(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i8m4_i8m1))) +vint8m1_t vget_i8m1(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i8m8_i8m1))) +vint8m1_t vget_i8m1(vint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i16m2_i16m1))) +vint16m1_t vget_i16m1(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i16m4_i16m1))) +vint16m1_t vget_i16m1(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i16m8_i16m1))) +vint16m1_t vget_i16m1(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i32m2_i32m1))) +vint32m1_t vget_i32m1(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i32m4_i32m1))) +vint32m1_t vget_i32m1(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i32m8_i32m1))) +vint32m1_t vget_i32m1(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i64m2_i64m1))) +vint64m1_t vget_i64m1(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i64m4_i64m1))) +vint64m1_t vget_i64m1(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i64m8_i64m1))) +vint64m1_t vget_i64m1(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u8m2_u8m1))) +vuint8m1_t vget_u8m1(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u8m4_u8m1))) +vuint8m1_t vget_u8m1(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u8m8_u8m1))) +vuint8m1_t vget_u8m1(vuint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u16m2_u16m1))) +vuint16m1_t vget_u16m1(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u16m4_u16m1))) +vuint16m1_t vget_u16m1(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u16m8_u16m1))) +vuint16m1_t vget_u16m1(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u32m2_u32m1))) +vuint32m1_t vget_u32m1(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u32m4_u32m1))) +vuint32m1_t vget_u32m1(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u32m8_u32m1))) +vuint32m1_t vget_u32m1(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u64m2_u64m1))) +vuint64m1_t vget_u64m1(vuint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u64m4_u64m1))) +vuint64m1_t vget_u64m1(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u64m8_u64m1))) +vuint64m1_t vget_u64m1(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i8m4_i8m2))) +vint8m2_t vget_i8m2(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i8m8_i8m2))) +vint8m2_t vget_i8m2(vint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i16m4_i16m2))) +vint16m2_t vget_i16m2(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i16m8_i16m2))) +vint16m2_t vget_i16m2(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i32m4_i32m2))) +vint32m2_t vget_i32m2(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i32m8_i32m2))) +vint32m2_t vget_i32m2(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i64m4_i64m2))) +vint64m2_t vget_i64m2(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i64m8_i64m2))) +vint64m2_t vget_i64m2(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u8m4_u8m2))) +vuint8m2_t vget_u8m2(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u8m8_u8m2))) +vuint8m2_t vget_u8m2(vuint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u16m4_u16m2))) +vuint16m2_t vget_u16m2(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u16m8_u16m2))) +vuint16m2_t vget_u16m2(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u32m4_u32m2))) +vuint32m2_t vget_u32m2(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u32m8_u32m2))) +vuint32m2_t vget_u32m2(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u64m4_u64m2))) +vuint64m2_t vget_u64m2(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u64m8_u64m2))) +vuint64m2_t vget_u64m2(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i8m8_i8m4))) +vint8m4_t vget_i8m4(vint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i16m8_i16m4))) +vint16m4_t vget_i16m4(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i32m8_i32m4))) +vint32m4_t vget_i32m4(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i64m8_i64m4))) +vint64m4_t vget_i64m4(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u8m8_u8m4))) +vuint8m4_t vget_u8m4(vuint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u16m8_u16m4))) +vuint16m4_t vget_u16m4(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u32m8_u32m4))) +vuint32m4_t vget_u32m4(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u64m8_u64m4))) +vuint64m4_t vget_u64m4(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i8m1_i8m2))) +vint8m2_t vset(vint8m2_t op0, size_t op1, vint8m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i16m1_i16m2))) +vint16m2_t vset(vint16m2_t op0, size_t op1, vint16m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i32m1_i32m2))) +vint32m2_t vset(vint32m2_t op0, size_t op1, vint32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i64m1_i64m2))) +vint64m2_t vset(vint64m2_t op0, size_t op1, vint64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u8m1_u8m2))) +vuint8m2_t vset(vuint8m2_t op0, size_t op1, vuint8m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u16m1_u16m2))) +vuint16m2_t vset(vuint16m2_t op0, size_t op1, vuint16m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u32m1_u32m2))) +vuint32m2_t vset(vuint32m2_t op0, size_t op1, vuint32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u64m1_u64m2))) +vuint64m2_t vset(vuint64m2_t op0, size_t op1, vuint64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i8m1_i8m4))) +vint8m4_t vset(vint8m4_t op0, size_t op1, vint8m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i8m2_i8m4))) +vint8m4_t vset(vint8m4_t op0, size_t op1, vint8m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i16m1_i16m4))) +vint16m4_t vset(vint16m4_t op0, size_t op1, vint16m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i16m2_i16m4))) +vint16m4_t vset(vint16m4_t op0, size_t op1, vint16m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i32m1_i32m4))) +vint32m4_t vset(vint32m4_t op0, size_t op1, vint32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i32m2_i32m4))) +vint32m4_t vset(vint32m4_t op0, size_t op1, vint32m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i64m1_i64m4))) +vint64m4_t vset(vint64m4_t op0, size_t op1, vint64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i64m2_i64m4))) +vint64m4_t vset(vint64m4_t op0, size_t op1, vint64m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u8m1_u8m4))) +vuint8m4_t vset(vuint8m4_t op0, size_t op1, vuint8m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u8m2_u8m4))) +vuint8m4_t vset(vuint8m4_t op0, size_t op1, vuint8m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u16m1_u16m4))) +vuint16m4_t vset(vuint16m4_t op0, size_t op1, vuint16m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u16m2_u16m4))) +vuint16m4_t vset(vuint16m4_t op0, size_t op1, vuint16m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u32m1_u32m4))) +vuint32m4_t vset(vuint32m4_t op0, size_t op1, vuint32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u32m2_u32m4))) +vuint32m4_t vset(vuint32m4_t op0, size_t op1, vuint32m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u64m1_u64m4))) +vuint64m4_t vset(vuint64m4_t op0, size_t op1, vuint64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u64m2_u64m4))) +vuint64m4_t vset(vuint64m4_t op0, size_t op1, vuint64m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i8m1_i8m8))) +vint8m8_t vset(vint8m8_t op0, size_t op1, vint8m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i8m2_i8m8))) +vint8m8_t vset(vint8m8_t op0, size_t op1, vint8m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i8m4_i8m8))) +vint8m8_t vset(vint8m8_t op0, size_t op1, vint8m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i16m1_i16m8))) +vint16m8_t vset(vint16m8_t op0, size_t op1, vint16m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i16m2_i16m8))) +vint16m8_t vset(vint16m8_t op0, size_t op1, vint16m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i16m4_i16m8))) +vint16m8_t vset(vint16m8_t op0, size_t op1, vint16m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i32m1_i32m8))) +vint32m8_t vset(vint32m8_t op0, size_t op1, vint32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i32m2_i32m8))) +vint32m8_t vset(vint32m8_t op0, size_t op1, vint32m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i32m4_i32m8))) +vint32m8_t vset(vint32m8_t op0, size_t op1, vint32m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i64m1_i64m8))) +vint64m8_t vset(vint64m8_t op0, size_t op1, vint64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i64m2_i64m8))) +vint64m8_t vset(vint64m8_t op0, size_t op1, vint64m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i64m4_i64m8))) +vint64m8_t vset(vint64m8_t op0, size_t op1, vint64m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u8m1_u8m8))) +vuint8m8_t vset(vuint8m8_t op0, size_t op1, vuint8m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u8m2_u8m8))) +vuint8m8_t vset(vuint8m8_t op0, size_t op1, vuint8m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u8m4_u8m8))) +vuint8m8_t vset(vuint8m8_t op0, size_t op1, vuint8m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u16m1_u16m8))) +vuint16m8_t vset(vuint16m8_t op0, size_t op1, vuint16m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u16m2_u16m8))) +vuint16m8_t vset(vuint16m8_t op0, size_t op1, vuint16m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u16m4_u16m8))) +vuint16m8_t vset(vuint16m8_t op0, size_t op1, vuint16m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u32m1_u32m8))) +vuint32m8_t vset(vuint32m8_t op0, size_t op1, vuint32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u32m2_u32m8))) +vuint32m8_t vset(vuint32m8_t op0, size_t op1, vuint32m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u32m4_u32m8))) +vuint32m8_t vset(vuint32m8_t op0, size_t op1, vuint32m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u64m1_u64m8))) +vuint64m8_t vset(vuint64m8_t op0, size_t op1, vuint64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u64m2_u64m8))) +vuint64m8_t vset(vuint64m8_t op0, size_t op1, vuint64m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u64m4_u64m8))) +vuint64m8_t vset(vuint64m8_t op0, size_t op1, vuint64m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m1))) +void vsoxei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m1_m))) +void vsoxei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m2))) +void vsoxei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m2_m))) +void vsoxei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m4))) +void vsoxei8(int8_t * op0, vuint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m4_m))) +void vsoxei8(vbool2_t op0, int8_t * op1, vuint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m8))) +void vsoxei8(int8_t * op0, vuint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m8_m))) +void vsoxei8(vbool1_t op0, int8_t * op1, vuint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8mf2))) +void vsoxei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8mf2_m))) +void vsoxei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8mf4))) +void vsoxei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8mf4_m))) +void vsoxei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8mf8))) +void vsoxei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8mf8_m))) +void vsoxei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m1))) +void vsoxei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m1_m))) +void vsoxei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m2))) +void vsoxei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m2_m))) +void vsoxei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m4))) +void vsoxei8(uint8_t * op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m4_m))) +void vsoxei8(vbool2_t op0, uint8_t * op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m8))) +void vsoxei8(uint8_t * op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m8_m))) +void vsoxei8(vbool1_t op0, uint8_t * op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8mf2))) +void vsoxei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8mf2_m))) +void vsoxei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8mf4))) +void vsoxei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8mf4_m))) +void vsoxei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8mf8))) +void vsoxei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8mf8_m))) +void vsoxei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8m1))) +void vsoxei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8m1_m))) +void vsoxei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8m2))) +void vsoxei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8m2_m))) +void vsoxei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8m4))) +void vsoxei16(int8_t * op0, vuint16m8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8m4_m))) +void vsoxei16(vbool2_t op0, int8_t * op1, vuint16m8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8mf2))) +void vsoxei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8mf2_m))) +void vsoxei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8mf4))) +void vsoxei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8mf4_m))) +void vsoxei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8mf8))) +void vsoxei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8mf8_m))) +void vsoxei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8m1))) +void vsoxei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8m1_m))) +void vsoxei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8m2))) +void vsoxei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8m2_m))) +void vsoxei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8m4))) +void vsoxei16(uint8_t * op0, vuint16m8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8m4_m))) +void vsoxei16(vbool2_t op0, uint8_t * op1, vuint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8mf2))) +void vsoxei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8mf2_m))) +void vsoxei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8mf4))) +void vsoxei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8mf4_m))) +void vsoxei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8mf8))) +void vsoxei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8mf8_m))) +void vsoxei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8m1))) +void vsoxei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8m1_m))) +void vsoxei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8m2))) +void vsoxei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8m2_m))) +void vsoxei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8mf2))) +void vsoxei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8mf2_m))) +void vsoxei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8mf4))) +void vsoxei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8mf4_m))) +void vsoxei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8mf8))) +void vsoxei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8mf8_m))) +void vsoxei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8m1))) +void vsoxei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8m1_m))) +void vsoxei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8m2))) +void vsoxei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8m2_m))) +void vsoxei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8mf2))) +void vsoxei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8mf2_m))) +void vsoxei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8mf4))) +void vsoxei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8mf4_m))) +void vsoxei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8mf8))) +void vsoxei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8mf8_m))) +void vsoxei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8m1))) +void vsoxei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8m1_m))) +void vsoxei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8mf2))) +void vsoxei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8mf2_m))) +void vsoxei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8mf4))) +void vsoxei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8mf4_m))) +void vsoxei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8mf8))) +void vsoxei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8mf8_m))) +void vsoxei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8m1))) +void vsoxei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8m1_m))) +void vsoxei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8mf2))) +void vsoxei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8mf2_m))) +void vsoxei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8mf4))) +void vsoxei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8mf4_m))) +void vsoxei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8mf8))) +void vsoxei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8mf8_m))) +void vsoxei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m1))) +void vsoxei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m1_m))) +void vsoxei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m2))) +void vsoxei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m2_m))) +void vsoxei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m4))) +void vsoxei8(int16_t * op0, vuint8m2_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m4_m))) +void vsoxei8(vbool4_t op0, int16_t * op1, vuint8m2_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m8))) +void vsoxei8(int16_t * op0, vuint8m4_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m8_m))) +void vsoxei8(vbool2_t op0, int16_t * op1, vuint8m4_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16mf2))) +void vsoxei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16mf2_m))) +void vsoxei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16mf4))) +void vsoxei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16mf4_m))) +void vsoxei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m1))) +void vsoxei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m1_m))) +void vsoxei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m2))) +void vsoxei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m2_m))) +void vsoxei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m4))) +void vsoxei8(uint16_t * op0, vuint8m2_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m4_m))) +void vsoxei8(vbool4_t op0, uint16_t * op1, vuint8m2_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m8))) +void vsoxei8(uint16_t * op0, vuint8m4_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m8_m))) +void vsoxei8(vbool2_t op0, uint16_t * op1, vuint8m4_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16mf2))) +void vsoxei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16mf2_m))) +void vsoxei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16mf4))) +void vsoxei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16mf4_m))) +void vsoxei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m1))) +void vsoxei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m1_m))) +void vsoxei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m2))) +void vsoxei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m2_m))) +void vsoxei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m4))) +void vsoxei16(int16_t * op0, vuint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m4_m))) +void vsoxei16(vbool4_t op0, int16_t * op1, vuint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m8))) +void vsoxei16(int16_t * op0, vuint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m8_m))) +void vsoxei16(vbool2_t op0, int16_t * op1, vuint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16mf2))) +void vsoxei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16mf2_m))) +void vsoxei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16mf4))) +void vsoxei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16mf4_m))) +void vsoxei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m1))) +void vsoxei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m1_m))) +void vsoxei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m2))) +void vsoxei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m2_m))) +void vsoxei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m4))) +void vsoxei16(uint16_t * op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m4_m))) +void vsoxei16(vbool4_t op0, uint16_t * op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m8))) +void vsoxei16(uint16_t * op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m8_m))) +void vsoxei16(vbool2_t op0, uint16_t * op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16mf2))) +void vsoxei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16mf2_m))) +void vsoxei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16mf4))) +void vsoxei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16mf4_m))) +void vsoxei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16m1))) +void vsoxei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16m1_m))) +void vsoxei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16m2))) +void vsoxei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16m2_m))) +void vsoxei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16m4))) +void vsoxei32(int16_t * op0, vuint32m8_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16m4_m))) +void vsoxei32(vbool4_t op0, int16_t * op1, vuint32m8_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16mf2))) +void vsoxei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16mf2_m))) +void vsoxei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16mf4))) +void vsoxei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16mf4_m))) +void vsoxei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16m1))) +void vsoxei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16m1_m))) +void vsoxei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16m2))) +void vsoxei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16m2_m))) +void vsoxei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16m4))) +void vsoxei32(uint16_t * op0, vuint32m8_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16m4_m))) +void vsoxei32(vbool4_t op0, uint16_t * op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16mf2))) +void vsoxei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16mf2_m))) +void vsoxei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16mf4))) +void vsoxei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16mf4_m))) +void vsoxei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16m1))) +void vsoxei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16m1_m))) +void vsoxei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16m2))) +void vsoxei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16m2_m))) +void vsoxei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16mf2))) +void vsoxei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16mf2_m))) +void vsoxei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16mf4))) +void vsoxei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16mf4_m))) +void vsoxei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16m1))) +void vsoxei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16m1_m))) +void vsoxei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16m2))) +void vsoxei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16m2_m))) +void vsoxei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16mf2))) +void vsoxei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16mf2_m))) +void vsoxei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16mf4))) +void vsoxei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16mf4_m))) +void vsoxei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m1))) +void vsoxei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m1_m))) +void vsoxei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m2))) +void vsoxei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m2_m))) +void vsoxei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m4))) +void vsoxei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m4_m))) +void vsoxei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m8))) +void vsoxei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m8_m))) +void vsoxei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32mf2))) +void vsoxei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32mf2_m))) +void vsoxei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m1))) +void vsoxei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m1_m))) +void vsoxei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m2))) +void vsoxei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m2_m))) +void vsoxei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m4))) +void vsoxei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m4_m))) +void vsoxei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m8))) +void vsoxei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m8_m))) +void vsoxei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32mf2))) +void vsoxei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32mf2_m))) +void vsoxei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m1))) +void vsoxei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m1_m))) +void vsoxei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m2))) +void vsoxei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m2_m))) +void vsoxei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m4))) +void vsoxei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m4_m))) +void vsoxei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m8))) +void vsoxei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m8_m))) +void vsoxei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32mf2))) +void vsoxei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32mf2_m))) +void vsoxei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m1))) +void vsoxei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m1_m))) +void vsoxei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m2))) +void vsoxei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m2_m))) +void vsoxei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m4))) +void vsoxei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m4_m))) +void vsoxei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m8))) +void vsoxei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m8_m))) +void vsoxei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32mf2))) +void vsoxei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32mf2_m))) +void vsoxei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m1))) +void vsoxei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m1_m))) +void vsoxei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m2))) +void vsoxei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m2_m))) +void vsoxei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m4))) +void vsoxei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m4_m))) +void vsoxei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m8))) +void vsoxei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m8_m))) +void vsoxei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32mf2))) +void vsoxei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32mf2_m))) +void vsoxei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m1))) +void vsoxei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m1_m))) +void vsoxei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m2))) +void vsoxei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m2_m))) +void vsoxei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m4))) +void vsoxei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m4_m))) +void vsoxei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m8))) +void vsoxei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m8_m))) +void vsoxei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32mf2))) +void vsoxei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32mf2_m))) +void vsoxei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32m1))) +void vsoxei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32m1_m))) +void vsoxei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32m2))) +void vsoxei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32m2_m))) +void vsoxei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32m4))) +void vsoxei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32m4_m))) +void vsoxei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32mf2))) +void vsoxei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32mf2_m))) +void vsoxei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32m1))) +void vsoxei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32m1_m))) +void vsoxei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32m2))) +void vsoxei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32m2_m))) +void vsoxei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32m4))) +void vsoxei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32m4_m))) +void vsoxei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32mf2))) +void vsoxei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32mf2_m))) +void vsoxei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m1))) +void vsoxei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m1_m))) +void vsoxei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m2))) +void vsoxei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m2_m))) +void vsoxei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m4))) +void vsoxei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m4_m))) +void vsoxei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m8))) +void vsoxei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m8_m))) +void vsoxei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m1))) +void vsoxei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m1_m))) +void vsoxei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m2))) +void vsoxei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m2_m))) +void vsoxei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m4))) +void vsoxei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m4_m))) +void vsoxei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m8))) +void vsoxei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m8_m))) +void vsoxei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m1))) +void vsoxei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m1_m))) +void vsoxei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m2))) +void vsoxei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m2_m))) +void vsoxei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m4))) +void vsoxei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m4_m))) +void vsoxei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m8))) +void vsoxei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m8_m))) +void vsoxei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m1))) +void vsoxei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m1_m))) +void vsoxei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m2))) +void vsoxei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m2_m))) +void vsoxei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m4))) +void vsoxei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m4_m))) +void vsoxei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m8))) +void vsoxei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m8_m))) +void vsoxei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m1))) +void vsoxei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m1_m))) +void vsoxei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m2))) +void vsoxei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m2_m))) +void vsoxei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m4))) +void vsoxei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m4_m))) +void vsoxei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m8))) +void vsoxei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m8_m))) +void vsoxei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m1))) +void vsoxei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m1_m))) +void vsoxei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m2))) +void vsoxei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m2_m))) +void vsoxei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m4))) +void vsoxei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m4_m))) +void vsoxei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m8))) +void vsoxei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m8_m))) +void vsoxei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m1))) +void vsoxei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m1_m))) +void vsoxei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m2))) +void vsoxei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m2_m))) +void vsoxei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m4))) +void vsoxei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m4_m))) +void vsoxei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m8))) +void vsoxei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m8_m))) +void vsoxei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m1))) +void vsoxei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m1_m))) +void vsoxei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m2))) +void vsoxei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m2_m))) +void vsoxei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m4))) +void vsoxei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m4_m))) +void vsoxei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m8))) +void vsoxei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m8_m))) +void vsoxei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_i16m1_m))) +vint16m1_t vle16ff(vbool16_t op0, vint16m1_t op1, const int16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_i16m2_m))) +vint16m2_t vle16ff(vbool8_t op0, vint16m2_t op1, const int16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_i16m4_m))) +vint16m4_t vle16ff(vbool4_t op0, vint16m4_t op1, const int16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_i16m8_m))) +vint16m8_t vle16ff(vbool2_t op0, vint16m8_t op1, const int16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_i16mf2_m))) +vint16mf2_t vle16ff(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_i16mf4_m))) +vint16mf4_t vle16ff(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_u16m1_m))) +vuint16m1_t vle16ff(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_u16m2_m))) +vuint16m2_t vle16ff(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_u16m4_m))) +vuint16m4_t vle16ff(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_u16m8_m))) +vuint16m8_t vle16ff(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_u16mf2_m))) +vuint16mf2_t vle16ff(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_u16mf4_m))) +vuint16mf4_t vle16ff(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_i32m1_m))) +vint32m1_t vle32ff(vbool32_t op0, vint32m1_t op1, const int32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_i32m2_m))) +vint32m2_t vle32ff(vbool16_t op0, vint32m2_t op1, const int32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_i32m4_m))) +vint32m4_t vle32ff(vbool8_t op0, vint32m4_t op1, const int32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_i32m8_m))) +vint32m8_t vle32ff(vbool4_t op0, vint32m8_t op1, const int32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_i32mf2_m))) +vint32mf2_t vle32ff(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_u32m1_m))) +vuint32m1_t vle32ff(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_u32m2_m))) +vuint32m2_t vle32ff(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_u32m4_m))) +vuint32m4_t vle32ff(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_u32m8_m))) +vuint32m8_t vle32ff(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_u32mf2_m))) +vuint32mf2_t vle32ff(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_i64m1_m))) +vint64m1_t vle64ff(vbool64_t op0, vint64m1_t op1, const int64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_i64m2_m))) +vint64m2_t vle64ff(vbool32_t op0, vint64m2_t op1, const int64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_i64m4_m))) +vint64m4_t vle64ff(vbool16_t op0, vint64m4_t op1, const int64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_i64m8_m))) +vint64m8_t vle64ff(vbool8_t op0, vint64m8_t op1, const int64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_u64m1_m))) +vuint64m1_t vle64ff(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_u64m2_m))) +vuint64m2_t vle64ff(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_u64m4_m))) +vuint64m4_t vle64ff(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_u64m8_m))) +vuint64m8_t vle64ff(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8m1_m))) +vint8m1_t vle8(vbool8_t op0, vint8m1_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8m2_m))) +vint8m2_t vle8(vbool4_t op0, vint8m2_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8m4_m))) +vint8m4_t vle8(vbool2_t op0, vint8m4_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8m8_m))) +vint8m8_t vle8(vbool1_t op0, vint8m8_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8mf2_m))) +vint8mf2_t vle8(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8mf4_m))) +vint8mf4_t vle8(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8mf8_m))) +vint8mf8_t vle8(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8m1_m))) +vuint8m1_t vle8(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8m2_m))) +vuint8m2_t vle8(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8m4_m))) +vuint8m4_t vle8(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8m8_m))) +vuint8m8_t vle8(vbool1_t op0, vuint8m8_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8mf2_m))) +vuint8mf2_t vle8(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8mf4_m))) +vuint8mf4_t vle8(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8mf8_m))) +vuint8mf8_t vle8(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8m1_m))) +vint8m1_t vle8ff(vbool8_t op0, vint8m1_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8m2_m))) +vint8m2_t vle8ff(vbool4_t op0, vint8m2_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8m4_m))) +vint8m4_t vle8ff(vbool2_t op0, vint8m4_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8m8_m))) +vint8m8_t vle8ff(vbool1_t op0, vint8m8_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8mf2_m))) +vint8mf2_t vle8ff(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8mf4_m))) +vint8mf4_t vle8ff(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8mf8_m))) +vint8mf8_t vle8ff(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8m1_m))) +vuint8m1_t vle8ff(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8m2_m))) +vuint8m2_t vle8ff(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8m4_m))) +vuint8m4_t vle8ff(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8m8_m))) +vuint8m8_t vle8ff(vbool1_t op0, vuint8m8_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8mf2_m))) +vuint8mf2_t vle8ff(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8mf4_m))) +vuint8mf4_t vle8ff(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8mf8_m))) +vuint8mf8_t vle8ff(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m1))) +vint8m1_t vneg(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m1_m))) +vint8m1_t vneg(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m2))) +vint8m2_t vneg(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m2_m))) +vint8m2_t vneg(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m4))) +vint8m4_t vneg(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m4_m))) +vint8m4_t vneg(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m8))) +vint8m8_t vneg(vint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m8_m))) +vint8m8_t vneg(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8mf2))) +vint8mf2_t vneg(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8mf2_m))) +vint8mf2_t vneg(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8mf4))) +vint8mf4_t vneg(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8mf4_m))) +vint8mf4_t vneg(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8mf8))) +vint8mf8_t vneg(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8mf8_m))) +vint8mf8_t vneg(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m1))) +vint16m1_t vneg(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m1_m))) +vint16m1_t vneg(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m2))) +vint16m2_t vneg(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m2_m))) +vint16m2_t vneg(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m4))) +vint16m4_t vneg(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m4_m))) +vint16m4_t vneg(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m8))) +vint16m8_t vneg(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m8_m))) +vint16m8_t vneg(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16mf2))) +vint16mf2_t vneg(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16mf2_m))) +vint16mf2_t vneg(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16mf4))) +vint16mf4_t vneg(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16mf4_m))) +vint16mf4_t vneg(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m1))) +vint32m1_t vneg(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m1_m))) +vint32m1_t vneg(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m2))) +vint32m2_t vneg(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m2_m))) +vint32m2_t vneg(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m4))) +vint32m4_t vneg(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m4_m))) +vint32m4_t vneg(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m8))) +vint32m8_t vneg(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m8_m))) +vint32m8_t vneg(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32mf2))) +vint32mf2_t vneg(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32mf2_m))) +vint32mf2_t vneg(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m1))) +vint64m1_t vneg(vint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m1_m))) +vint64m1_t vneg(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m2))) +vint64m2_t vneg(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m2_m))) +vint64m2_t vneg(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m4))) +vint64m4_t vneg(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m4_m))) +vint64m4_t vneg(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m8))) +vint64m8_t vneg(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m8_m))) +vint64m8_t vneg(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m1))) +vint8m1_t vnot(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m1_m))) +vint8m1_t vnot(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m2))) +vint8m2_t vnot(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m2_m))) +vint8m2_t vnot(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m4))) +vint8m4_t vnot(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m4_m))) +vint8m4_t vnot(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m8))) +vint8m8_t vnot(vint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m8_m))) +vint8m8_t vnot(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8mf2))) +vint8mf2_t vnot(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8mf2_m))) +vint8mf2_t vnot(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8mf4))) +vint8mf4_t vnot(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8mf4_m))) +vint8mf4_t vnot(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8mf8))) +vint8mf8_t vnot(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8mf8_m))) +vint8mf8_t vnot(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m1))) +vint16m1_t vnot(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m1_m))) +vint16m1_t vnot(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m2))) +vint16m2_t vnot(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m2_m))) +vint16m2_t vnot(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m4))) +vint16m4_t vnot(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m4_m))) +vint16m4_t vnot(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m8))) +vint16m8_t vnot(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m8_m))) +vint16m8_t vnot(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16mf2))) +vint16mf2_t vnot(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16mf2_m))) +vint16mf2_t vnot(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16mf4))) +vint16mf4_t vnot(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16mf4_m))) +vint16mf4_t vnot(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m1))) +vint32m1_t vnot(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m1_m))) +vint32m1_t vnot(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m2))) +vint32m2_t vnot(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m2_m))) +vint32m2_t vnot(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m4))) +vint32m4_t vnot(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m4_m))) +vint32m4_t vnot(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m8))) +vint32m8_t vnot(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m8_m))) +vint32m8_t vnot(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32mf2))) +vint32mf2_t vnot(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32mf2_m))) +vint32mf2_t vnot(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m1))) +vint64m1_t vnot(vint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m1_m))) +vint64m1_t vnot(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m2))) +vint64m2_t vnot(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m2_m))) +vint64m2_t vnot(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m4))) +vint64m4_t vnot(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m4_m))) +vint64m4_t vnot(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m8))) +vint64m8_t vnot(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m8_m))) +vint64m8_t vnot(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8m1_m))) +vint8m1_t vlse8(vbool8_t op0, vint8m1_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8m2_m))) +vint8m2_t vlse8(vbool4_t op0, vint8m2_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8m4_m))) +vint8m4_t vlse8(vbool2_t op0, vint8m4_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8m8_m))) +vint8m8_t vlse8(vbool1_t op0, vint8m8_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8mf2_m))) +vint8mf2_t vlse8(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8mf4_m))) +vint8mf4_t vlse8(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8mf8_m))) +vint8mf8_t vlse8(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m1))) +vuint8m1_t vnot(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m1_m))) +vuint8m1_t vnot(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m2))) +vuint8m2_t vnot(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m2_m))) +vuint8m2_t vnot(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m4))) +vuint8m4_t vnot(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m4_m))) +vuint8m4_t vnot(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m8))) +vuint8m8_t vnot(vuint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m8_m))) +vuint8m8_t vnot(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8mf2))) +vuint8mf2_t vnot(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8mf2_m))) +vuint8mf2_t vnot(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8mf4))) +vuint8mf4_t vnot(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8mf4_m))) +vuint8mf4_t vnot(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8mf8))) +vuint8mf8_t vnot(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8mf8_m))) +vuint8mf8_t vnot(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m1))) +vuint16m1_t vnot(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m1_m))) +vuint16m1_t vnot(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m2))) +vuint16m2_t vnot(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m2_m))) +vuint16m2_t vnot(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m4))) +vuint16m4_t vnot(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m4_m))) +vuint16m4_t vnot(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m8))) +vuint16m8_t vnot(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m8_m))) +vuint16m8_t vnot(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16mf2))) +vuint16mf2_t vnot(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16mf2_m))) +vuint16mf2_t vnot(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16mf4))) +vuint16mf4_t vnot(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16mf4_m))) +vuint16mf4_t vnot(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m1))) +vuint32m1_t vnot(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m1_m))) +vuint32m1_t vnot(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m2))) +vuint32m2_t vnot(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m2_m))) +vuint32m2_t vnot(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m4))) +vuint32m4_t vnot(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m4_m))) +vuint32m4_t vnot(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m8))) +vuint32m8_t vnot(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m8_m))) +vuint32m8_t vnot(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32mf2))) +vuint32mf2_t vnot(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32mf2_m))) +vuint32mf2_t vnot(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m1))) +vuint64m1_t vnot(vuint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m1_m))) +vuint64m1_t vnot(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m2))) +vuint64m2_t vnot(vuint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m2_m))) +vuint64m2_t vnot(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m4))) +vuint64m4_t vnot(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m4_m))) +vuint64m4_t vnot(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m8))) +vuint64m8_t vnot(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m8_m))) +vuint64m8_t vnot(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b8))) +vbool8_t vmmv(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b4))) +vbool4_t vmmv(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b2))) +vbool2_t vmmv(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b1))) +vbool1_t vmmv(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b16))) +vbool16_t vmmv(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b32))) +vbool32_t vmmv(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b64))) +vbool64_t vmmv(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m2))) +vuint16m2_t vwcvtu_x(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m2_m))) +vuint16m2_t vwcvtu_x(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m4))) +vuint16m4_t vwcvtu_x(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m4_m))) +vuint16m4_t vwcvtu_x(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m8))) +vuint16m8_t vwcvtu_x(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m8_m))) +vuint16m8_t vwcvtu_x(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m1))) +vuint16m1_t vwcvtu_x(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m1_m))) +vuint16m1_t vwcvtu_x(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16mf2))) +vuint16mf2_t vwcvtu_x(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16mf2_m))) +vuint16mf2_t vwcvtu_x(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16mf4))) +vuint16mf4_t vwcvtu_x(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16mf4_m))) +vuint16mf4_t vwcvtu_x(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m2))) +vuint32m2_t vwcvtu_x(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m2_m))) +vuint32m2_t vwcvtu_x(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m4))) +vuint32m4_t vwcvtu_x(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m4_m))) +vuint32m4_t vwcvtu_x(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m8))) +vuint32m8_t vwcvtu_x(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m8_m))) +vuint32m8_t vwcvtu_x(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m1))) +vuint32m1_t vwcvtu_x(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m1_m))) +vuint32m1_t vwcvtu_x(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32mf2))) +vuint32mf2_t vwcvtu_x(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32mf2_m))) +vuint32mf2_t vwcvtu_x(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m2))) +vuint64m2_t vwcvtu_x(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m2_m))) +vuint64m2_t vwcvtu_x(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m4))) +vuint64m4_t vwcvtu_x(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m4_m))) +vuint64m4_t vwcvtu_x(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m8))) +vuint64m8_t vwcvtu_x(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m8_m))) +vuint64m8_t vwcvtu_x(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m1))) +vuint64m1_t vwcvtu_x(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m1_m))) +vuint64m1_t vwcvtu_x(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8m1))) +vint8m1_t vncvt_x(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8m1_m))) +vint8m1_t vncvt_x(vbool8_t op0, vint8m1_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8m2))) +vint8m2_t vncvt_x(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8m2_m))) +vint8m2_t vncvt_x(vbool4_t op0, vint8m2_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8m4))) +vint8m4_t vncvt_x(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8m4_m))) +vint8m4_t vncvt_x(vbool2_t op0, vint8m4_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8mf2))) +vint8mf2_t vncvt_x(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8mf2_m))) +vint8mf2_t vncvt_x(vbool16_t op0, vint8mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8mf4))) +vint8mf4_t vncvt_x(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8mf4_m))) +vint8mf4_t vncvt_x(vbool32_t op0, vint8mf4_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8mf8))) +vint8mf8_t vncvt_x(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8mf8_m))) +vint8mf8_t vncvt_x(vbool64_t op0, vint8mf8_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16m1))) +vint16m1_t vncvt_x(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16m1_m))) +vint16m1_t vncvt_x(vbool16_t op0, vint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16m2))) +vint16m2_t vncvt_x(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16m2_m))) +vint16m2_t vncvt_x(vbool8_t op0, vint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16m4))) +vint16m4_t vncvt_x(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16m4_m))) +vint16m4_t vncvt_x(vbool4_t op0, vint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16mf2))) +vint16mf2_t vncvt_x(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16mf2_m))) +vint16mf2_t vncvt_x(vbool32_t op0, vint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16mf4))) +vint16mf4_t vncvt_x(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16mf4_m))) +vint16mf4_t vncvt_x(vbool64_t op0, vint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32m1))) +vint32m1_t vncvt_x(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32m1_m))) +vint32m1_t vncvt_x(vbool32_t op0, vint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32m2))) +vint32m2_t vncvt_x(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32m2_m))) +vint32m2_t vncvt_x(vbool16_t op0, vint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32m4))) +vint32m4_t vncvt_x(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32m4_m))) +vint32m4_t vncvt_x(vbool8_t op0, vint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32mf2))) +vint32mf2_t vncvt_x(vint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32mf2_m))) +vint32mf2_t vncvt_x(vbool64_t op0, vint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_i16m1_m))) +vint16m1_t vle16(vbool16_t op0, vint16m1_t op1, const int16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_i16m2_m))) +vint16m2_t vle16(vbool8_t op0, vint16m2_t op1, const int16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_i16m4_m))) +vint16m4_t vle16(vbool4_t op0, vint16m4_t op1, const int16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_i16m8_m))) +vint16m8_t vle16(vbool2_t op0, vint16m8_t op1, const int16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_i16mf2_m))) +vint16mf2_t vle16(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_i16mf4_m))) +vint16mf4_t vle16(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_u16m1_m))) +vuint16m1_t vle16(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_u16m2_m))) +vuint16m2_t vle16(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_u16m4_m))) +vuint16m4_t vle16(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_u16m8_m))) +vuint16m8_t vle16(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_u16mf2_m))) +vuint16mf2_t vle16(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_u16mf4_m))) +vuint16mf4_t vle16(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_i32m1_m))) +vint32m1_t vle32(vbool32_t op0, vint32m1_t op1, const int32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_i32m2_m))) +vint32m2_t vle32(vbool16_t op0, vint32m2_t op1, const int32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_i32m4_m))) +vint32m4_t vle32(vbool8_t op0, vint32m4_t op1, const int32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_i32m8_m))) +vint32m8_t vle32(vbool4_t op0, vint32m8_t op1, const int32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_i32mf2_m))) +vint32mf2_t vle32(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_u32m1_m))) +vuint32m1_t vle32(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_u32m2_m))) +vuint32m2_t vle32(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_u32m4_m))) +vuint32m4_t vle32(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_u32m8_m))) +vuint32m8_t vle32(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_u32mf2_m))) +vuint32mf2_t vle32(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8m1_m))) +vuint8m1_t vlse8(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8m2_m))) +vuint8m2_t vlse8(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8m4_m))) +vuint8m4_t vlse8(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8m8_m))) +vuint8m8_t vlse8(vbool1_t op0, vuint8m8_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8mf2_m))) +vuint8mf2_t vlse8(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8mf4_m))) +vuint8mf4_t vlse8(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8mf8_m))) +vuint8mf8_t vlse8(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_i64m1_m))) +vint64m1_t vle64(vbool64_t op0, vint64m1_t op1, const int64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_i64m2_m))) +vint64m2_t vle64(vbool32_t op0, vint64m2_t op1, const int64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_i64m4_m))) +vint64m4_t vle64(vbool16_t op0, vint64m4_t op1, const int64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_i64m8_m))) +vint64m8_t vle64(vbool8_t op0, vint64m8_t op1, const int64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_u64m1_m))) +vuint64m1_t vle64(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_u64m2_m))) +vuint64m2_t vle64(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_u64m4_m))) +vuint64m4_t vle64(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_u64m8_m))) +vuint64m8_t vle64(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m1))) +void vse16(int16_t * op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m1_m))) +void vse16(vbool16_t op0, int16_t * op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m2))) +void vse16(int16_t * op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m2_m))) +void vse16(vbool8_t op0, int16_t * op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m4))) +void vse16(int16_t * op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m4_m))) +void vse16(vbool4_t op0, int16_t * op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m8))) +void vse16(int16_t * op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m8_m))) +void vse16(vbool2_t op0, int16_t * op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16mf2))) +void vse16(int16_t * op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16mf2_m))) +void vse16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16mf4))) +void vse16(int16_t * op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16mf4_m))) +void vse16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m1))) +void vse16(uint16_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m1_m))) +void vse16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m2))) +void vse16(uint16_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m2_m))) +void vse16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m4))) +void vse16(uint16_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m4_m))) +void vse16(vbool4_t op0, uint16_t * op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m8))) +void vse16(uint16_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m8_m))) +void vse16(vbool2_t op0, uint16_t * op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16mf2))) +void vse16(uint16_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16mf2_m))) +void vse16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16mf4))) +void vse16(uint16_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16mf4_m))) +void vse16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m1))) +void vse32(int32_t * op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m1_m))) +void vse32(vbool32_t op0, int32_t * op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m2))) +void vse32(int32_t * op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m2_m))) +void vse32(vbool16_t op0, int32_t * op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m4))) +void vse32(int32_t * op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m4_m))) +void vse32(vbool8_t op0, int32_t * op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m8))) +void vse32(int32_t * op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m8_m))) +void vse32(vbool4_t op0, int32_t * op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32mf2))) +void vse32(int32_t * op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32mf2_m))) +void vse32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m1))) +void vse32(uint32_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m1_m))) +void vse32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m2))) +void vse32(uint32_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m2_m))) +void vse32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m4))) +void vse32(uint32_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m4_m))) +void vse32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m8))) +void vse32(uint32_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m8_m))) +void vse32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32mf2))) +void vse32(uint32_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32mf2_m))) +void vse32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m1))) +vint8m1_t vluxei8(const int8_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m1_m))) +vint8m1_t vluxei8(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m2))) +vint8m2_t vluxei8(const int8_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m2_m))) +vint8m2_t vluxei8(vbool4_t op0, vint8m2_t op1, const int8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m4))) +vint8m4_t vluxei8(const int8_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m4_m))) +vint8m4_t vluxei8(vbool2_t op0, vint8m4_t op1, const int8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m8))) +vint8m8_t vluxei8(const int8_t * op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m8_m))) +vint8m8_t vluxei8(vbool1_t op0, vint8m8_t op1, const int8_t * op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8mf2))) +vint8mf2_t vluxei8(const int8_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8mf2_m))) +vint8mf2_t vluxei8(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8mf4))) +vint8mf4_t vluxei8(const int8_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8mf4_m))) +vint8mf4_t vluxei8(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8mf8))) +vint8mf8_t vluxei8(const int8_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8mf8_m))) +vint8mf8_t vluxei8(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m1))) +void vse64(int64_t * op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m1_m))) +void vse64(vbool64_t op0, int64_t * op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m2))) +void vse64(int64_t * op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m2_m))) +void vse64(vbool32_t op0, int64_t * op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m4))) +void vse64(int64_t * op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m4_m))) +void vse64(vbool16_t op0, int64_t * op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m8))) +void vse64(int64_t * op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m8_m))) +void vse64(vbool8_t op0, int64_t * op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m1))) +void vse64(uint64_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m1_m))) +void vse64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m2))) +void vse64(uint64_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m2_m))) +void vse64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m4))) +void vse64(uint64_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m4_m))) +void vse64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m8))) +void vse64(uint64_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m8_m))) +void vse64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_i16m1_m))) +vint16m1_t vlse16(vbool16_t op0, vint16m1_t op1, const int16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_i16m2_m))) +vint16m2_t vlse16(vbool8_t op0, vint16m2_t op1, const int16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_i16m4_m))) +vint16m4_t vlse16(vbool4_t op0, vint16m4_t op1, const int16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_i16m8_m))) +vint16m8_t vlse16(vbool2_t op0, vint16m8_t op1, const int16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_i16mf2_m))) +vint16mf2_t vlse16(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_i16mf4_m))) +vint16mf4_t vlse16(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_u16m1_m))) +vuint16m1_t vlse16(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_u16m2_m))) +vuint16m2_t vlse16(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_u16m4_m))) +vuint16m4_t vlse16(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_u16m8_m))) +vuint16m8_t vlse16(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_u16mf2_m))) +vuint16mf2_t vlse16(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_u16mf4_m))) +vuint16mf4_t vlse16(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_i32m1_m))) +vint32m1_t vlse32(vbool32_t op0, vint32m1_t op1, const int32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_i32m2_m))) +vint32m2_t vlse32(vbool16_t op0, vint32m2_t op1, const int32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_i32m4_m))) +vint32m4_t vlse32(vbool8_t op0, vint32m4_t op1, const int32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_i32m8_m))) +vint32m8_t vlse32(vbool4_t op0, vint32m8_t op1, const int32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_i32mf2_m))) +vint32mf2_t vlse32(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_u32m1_m))) +vuint32m1_t vlse32(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_u32m2_m))) +vuint32m2_t vlse32(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_u32m4_m))) +vuint32m4_t vlse32(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_u32m8_m))) +vuint32m8_t vlse32(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_u32mf2_m))) +vuint32mf2_t vlse32(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_i64m1_m))) +vint64m1_t vlse64(vbool64_t op0, vint64m1_t op1, const int64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_i64m2_m))) +vint64m2_t vlse64(vbool32_t op0, vint64m2_t op1, const int64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_i64m4_m))) +vint64m4_t vlse64(vbool16_t op0, vint64m4_t op1, const int64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_i64m8_m))) +vint64m8_t vlse64(vbool8_t op0, vint64m8_t op1, const int64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b8))) +long vfirst(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b8_m))) +long vfirst(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b4))) +long vfirst(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b4_m))) +long vfirst(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b2))) +long vfirst(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b2_m))) +long vfirst(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b1))) +long vfirst(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b1_m))) +long vfirst(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b16))) +long vfirst(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b16_m))) +long vfirst(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b32))) +long vfirst(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b32_m))) +long vfirst(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b64))) +long vfirst(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b64_m))) +long vfirst(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf2_i8m1))) +vint8m1_t vlmul_ext_i8m1(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf4_i8m1))) +vint8m1_t vlmul_ext_i8m1(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf8_i8m1))) +vint8m1_t vlmul_ext_i8m1(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf2_i16m1))) +vint16m1_t vlmul_ext_i16m1(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf4_i16m1))) +vint16m1_t vlmul_ext_i16m1(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32mf2_i32m1))) +vint32m1_t vlmul_ext_i32m1(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf4_i8mf2))) +vint8mf2_t vlmul_ext_i8mf2(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf8_i8mf2))) +vint8mf2_t vlmul_ext_i8mf2(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf4_i16mf2))) +vint16mf2_t vlmul_ext_i16mf2(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf8_i8mf4))) +vint8mf4_t vlmul_ext_i8mf4(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8m1_i8m2))) +vint8m2_t vlmul_ext_i8m2(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf2_i8m2))) +vint8m2_t vlmul_ext_i8m2(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf4_i8m2))) +vint8m2_t vlmul_ext_i8m2(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf8_i8m2))) +vint8m2_t vlmul_ext_i8m2(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16m1_i16m2))) +vint16m2_t vlmul_ext_i16m2(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf2_i16m2))) +vint16m2_t vlmul_ext_i16m2(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf4_i16m2))) +vint16m2_t vlmul_ext_i16m2(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32m1_i32m2))) +vint32m2_t vlmul_ext_i32m2(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32mf2_i32m2))) +vint32m2_t vlmul_ext_i32m2(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i64m1_i64m2))) +vint64m2_t vlmul_ext_i64m2(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8m1_i8m4))) +vint8m4_t vlmul_ext_i8m4(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8m2_i8m4))) +vint8m4_t vlmul_ext_i8m4(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf2_i8m4))) +vint8m4_t vlmul_ext_i8m4(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf4_i8m4))) +vint8m4_t vlmul_ext_i8m4(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf8_i8m4))) +vint8m4_t vlmul_ext_i8m4(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16m1_i16m4))) +vint16m4_t vlmul_ext_i16m4(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16m2_i16m4))) +vint16m4_t vlmul_ext_i16m4(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf2_i16m4))) +vint16m4_t vlmul_ext_i16m4(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf4_i16m4))) +vint16m4_t vlmul_ext_i16m4(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32m1_i32m4))) +vint32m4_t vlmul_ext_i32m4(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32m2_i32m4))) +vint32m4_t vlmul_ext_i32m4(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32mf2_i32m4))) +vint32m4_t vlmul_ext_i32m4(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i64m1_i64m4))) +vint64m4_t vlmul_ext_i64m4(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i64m2_i64m4))) +vint64m4_t vlmul_ext_i64m4(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8m1_i8m8))) +vint8m8_t vlmul_ext_i8m8(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8m2_i8m8))) +vint8m8_t vlmul_ext_i8m8(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8m4_i8m8))) +vint8m8_t vlmul_ext_i8m8(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf2_i8m8))) +vint8m8_t vlmul_ext_i8m8(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf4_i8m8))) +vint8m8_t vlmul_ext_i8m8(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf8_i8m8))) +vint8m8_t vlmul_ext_i8m8(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16m1_i16m8))) +vint16m8_t vlmul_ext_i16m8(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16m2_i16m8))) +vint16m8_t vlmul_ext_i16m8(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16m4_i16m8))) +vint16m8_t vlmul_ext_i16m8(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf2_i16m8))) +vint16m8_t vlmul_ext_i16m8(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf4_i16m8))) +vint16m8_t vlmul_ext_i16m8(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32m1_i32m8))) +vint32m8_t vlmul_ext_i32m8(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32m2_i32m8))) +vint32m8_t vlmul_ext_i32m8(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32m4_i32m8))) +vint32m8_t vlmul_ext_i32m8(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32mf2_i32m8))) +vint32m8_t vlmul_ext_i32m8(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i64m1_i64m8))) +vint64m8_t vlmul_ext_i64m8(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i64m2_i64m8))) +vint64m8_t vlmul_ext_i64m8(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i64m4_i64m8))) +vint64m8_t vlmul_ext_i64m8(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf2_u8m1))) +vuint8m1_t vlmul_ext_u8m1(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf4_u8m1))) +vuint8m1_t vlmul_ext_u8m1(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf8_u8m1))) +vuint8m1_t vlmul_ext_u8m1(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf2_u16m1))) +vuint16m1_t vlmul_ext_u16m1(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf4_u16m1))) +vuint16m1_t vlmul_ext_u16m1(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32mf2_u32m1))) +vuint32m1_t vlmul_ext_u32m1(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf4_u8mf2))) +vuint8mf2_t vlmul_ext_u8mf2(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf8_u8mf2))) +vuint8mf2_t vlmul_ext_u8mf2(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf4_u16mf2))) +vuint16mf2_t vlmul_ext_u16mf2(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf8_u8mf4))) +vuint8mf4_t vlmul_ext_u8mf4(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8m1_u8m2))) +vuint8m2_t vlmul_ext_u8m2(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf2_u8m2))) +vuint8m2_t vlmul_ext_u8m2(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf4_u8m2))) +vuint8m2_t vlmul_ext_u8m2(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf8_u8m2))) +vuint8m2_t vlmul_ext_u8m2(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16m1_u16m2))) +vuint16m2_t vlmul_ext_u16m2(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf2_u16m2))) +vuint16m2_t vlmul_ext_u16m2(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf4_u16m2))) +vuint16m2_t vlmul_ext_u16m2(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32m1_u32m2))) +vuint32m2_t vlmul_ext_u32m2(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32mf2_u32m2))) +vuint32m2_t vlmul_ext_u32m2(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u64m1_u64m2))) +vuint64m2_t vlmul_ext_u64m2(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8m1_u8m4))) +vuint8m4_t vlmul_ext_u8m4(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8m2_u8m4))) +vuint8m4_t vlmul_ext_u8m4(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf2_u8m4))) +vuint8m4_t vlmul_ext_u8m4(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf4_u8m4))) +vuint8m4_t vlmul_ext_u8m4(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf8_u8m4))) +vuint8m4_t vlmul_ext_u8m4(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16m1_u16m4))) +vuint16m4_t vlmul_ext_u16m4(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16m2_u16m4))) +vuint16m4_t vlmul_ext_u16m4(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf2_u16m4))) +vuint16m4_t vlmul_ext_u16m4(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf4_u16m4))) +vuint16m4_t vlmul_ext_u16m4(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32m1_u32m4))) +vuint32m4_t vlmul_ext_u32m4(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32m2_u32m4))) +vuint32m4_t vlmul_ext_u32m4(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32mf2_u32m4))) +vuint32m4_t vlmul_ext_u32m4(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u64m1_u64m4))) +vuint64m4_t vlmul_ext_u64m4(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u64m2_u64m4))) +vuint64m4_t vlmul_ext_u64m4(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8m1_u8m8))) +vuint8m8_t vlmul_ext_u8m8(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8m2_u8m8))) +vuint8m8_t vlmul_ext_u8m8(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8m4_u8m8))) +vuint8m8_t vlmul_ext_u8m8(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf2_u8m8))) +vuint8m8_t vlmul_ext_u8m8(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf4_u8m8))) +vuint8m8_t vlmul_ext_u8m8(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf8_u8m8))) +vuint8m8_t vlmul_ext_u8m8(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16m1_u16m8))) +vuint16m8_t vlmul_ext_u16m8(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16m2_u16m8))) +vuint16m8_t vlmul_ext_u16m8(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16m4_u16m8))) +vuint16m8_t vlmul_ext_u16m8(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf2_u16m8))) +vuint16m8_t vlmul_ext_u16m8(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf4_u16m8))) +vuint16m8_t vlmul_ext_u16m8(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32m1_u32m8))) +vuint32m8_t vlmul_ext_u32m8(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32m2_u32m8))) +vuint32m8_t vlmul_ext_u32m8(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32m4_u32m8))) +vuint32m8_t vlmul_ext_u32m8(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32mf2_u32m8))) +vuint32m8_t vlmul_ext_u32m8(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u64m1_u64m8))) +vuint64m8_t vlmul_ext_u64m8(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u64m2_u64m8))) +vuint64m8_t vlmul_ext_u64m8(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u64m4_u64m8))) +vuint64m8_t vlmul_ext_u64m8(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m1_i8mf2))) +vint8mf2_t vlmul_trunc_i8mf2(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m2_i8mf2))) +vint8mf2_t vlmul_trunc_i8mf2(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m4_i8mf2))) +vint8mf2_t vlmul_trunc_i8mf2(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m8_i8mf2))) +vint8mf2_t vlmul_trunc_i8mf2(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m1_i16mf2))) +vint16mf2_t vlmul_trunc_i16mf2(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m2_i16mf2))) +vint16mf2_t vlmul_trunc_i16mf2(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m4_i16mf2))) +vint16mf2_t vlmul_trunc_i16mf2(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m8_i16mf2))) +vint16mf2_t vlmul_trunc_i16mf2(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m1_i32mf2))) +vint32mf2_t vlmul_trunc_i32mf2(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m2_i32mf2))) +vint32mf2_t vlmul_trunc_i32mf2(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m4_i32mf2))) +vint32mf2_t vlmul_trunc_i32mf2(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m8_i32mf2))) +vint32mf2_t vlmul_trunc_i32mf2(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m1_i8mf4))) +vint8mf4_t vlmul_trunc_i8mf4(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m2_i8mf4))) +vint8mf4_t vlmul_trunc_i8mf4(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m4_i8mf4))) +vint8mf4_t vlmul_trunc_i8mf4(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m8_i8mf4))) +vint8mf4_t vlmul_trunc_i8mf4(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8mf2_i8mf4))) +vint8mf4_t vlmul_trunc_i8mf4(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m1_i16mf4))) +vint16mf4_t vlmul_trunc_i16mf4(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m2_i16mf4))) +vint16mf4_t vlmul_trunc_i16mf4(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m4_i16mf4))) +vint16mf4_t vlmul_trunc_i16mf4(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m8_i16mf4))) +vint16mf4_t vlmul_trunc_i16mf4(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16mf2_i16mf4))) +vint16mf4_t vlmul_trunc_i16mf4(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m1_i8mf8))) +vint8mf8_t vlmul_trunc_i8mf8(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m2_i8mf8))) +vint8mf8_t vlmul_trunc_i8mf8(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m4_i8mf8))) +vint8mf8_t vlmul_trunc_i8mf8(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m8_i8mf8))) +vint8mf8_t vlmul_trunc_i8mf8(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8mf2_i8mf8))) +vint8mf8_t vlmul_trunc_i8mf8(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8mf4_i8mf8))) +vint8mf8_t vlmul_trunc_i8mf8(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m2_i8m1))) +vint8m1_t vlmul_trunc_i8m1(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m4_i8m1))) +vint8m1_t vlmul_trunc_i8m1(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m8_i8m1))) +vint8m1_t vlmul_trunc_i8m1(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m2_i16m1))) +vint16m1_t vlmul_trunc_i16m1(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m4_i16m1))) +vint16m1_t vlmul_trunc_i16m1(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m8_i16m1))) +vint16m1_t vlmul_trunc_i16m1(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m2_i32m1))) +vint32m1_t vlmul_trunc_i32m1(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m4_i32m1))) +vint32m1_t vlmul_trunc_i32m1(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m8_i32m1))) +vint32m1_t vlmul_trunc_i32m1(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i64m2_i64m1))) +vint64m1_t vlmul_trunc_i64m1(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i64m4_i64m1))) +vint64m1_t vlmul_trunc_i64m1(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i64m8_i64m1))) +vint64m1_t vlmul_trunc_i64m1(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m4_i8m2))) +vint8m2_t vlmul_trunc_i8m2(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m8_i8m2))) +vint8m2_t vlmul_trunc_i8m2(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m4_i16m2))) +vint16m2_t vlmul_trunc_i16m2(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m8_i16m2))) +vint16m2_t vlmul_trunc_i16m2(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m4_i32m2))) +vint32m2_t vlmul_trunc_i32m2(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m8_i32m2))) +vint32m2_t vlmul_trunc_i32m2(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i64m4_i64m2))) +vint64m2_t vlmul_trunc_i64m2(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i64m8_i64m2))) +vint64m2_t vlmul_trunc_i64m2(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m8_i8m4))) +vint8m4_t vlmul_trunc_i8m4(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m8_i16m4))) +vint16m4_t vlmul_trunc_i16m4(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m8_i32m4))) +vint32m4_t vlmul_trunc_i32m4(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i64m8_i64m4))) +vint64m4_t vlmul_trunc_i64m4(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m1_u8mf2))) +vuint8mf2_t vlmul_trunc_u8mf2(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m2_u8mf2))) +vuint8mf2_t vlmul_trunc_u8mf2(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m4_u8mf2))) +vuint8mf2_t vlmul_trunc_u8mf2(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m8_u8mf2))) +vuint8mf2_t vlmul_trunc_u8mf2(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m1_u16mf2))) +vuint16mf2_t vlmul_trunc_u16mf2(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m2_u16mf2))) +vuint16mf2_t vlmul_trunc_u16mf2(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m4_u16mf2))) +vuint16mf2_t vlmul_trunc_u16mf2(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m8_u16mf2))) +vuint16mf2_t vlmul_trunc_u16mf2(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m1_u32mf2))) +vuint32mf2_t vlmul_trunc_u32mf2(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m2_u32mf2))) +vuint32mf2_t vlmul_trunc_u32mf2(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m4_u32mf2))) +vuint32mf2_t vlmul_trunc_u32mf2(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m8_u32mf2))) +vuint32mf2_t vlmul_trunc_u32mf2(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m1_u8mf4))) +vuint8mf4_t vlmul_trunc_u8mf4(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m2_u8mf4))) +vuint8mf4_t vlmul_trunc_u8mf4(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m4_u8mf4))) +vuint8mf4_t vlmul_trunc_u8mf4(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m8_u8mf4))) +vuint8mf4_t vlmul_trunc_u8mf4(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8mf2_u8mf4))) +vuint8mf4_t vlmul_trunc_u8mf4(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m1_u16mf4))) +vuint16mf4_t vlmul_trunc_u16mf4(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m2_u16mf4))) +vuint16mf4_t vlmul_trunc_u16mf4(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m4_u16mf4))) +vuint16mf4_t vlmul_trunc_u16mf4(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m8_u16mf4))) +vuint16mf4_t vlmul_trunc_u16mf4(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16mf2_u16mf4))) +vuint16mf4_t vlmul_trunc_u16mf4(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m1_u8mf8))) +vuint8mf8_t vlmul_trunc_u8mf8(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m2_u8mf8))) +vuint8mf8_t vlmul_trunc_u8mf8(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m4_u8mf8))) +vuint8mf8_t vlmul_trunc_u8mf8(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m8_u8mf8))) +vuint8mf8_t vlmul_trunc_u8mf8(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8mf2_u8mf8))) +vuint8mf8_t vlmul_trunc_u8mf8(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8mf4_u8mf8))) +vuint8mf8_t vlmul_trunc_u8mf8(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m2_u8m1))) +vuint8m1_t vlmul_trunc_u8m1(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m4_u8m1))) +vuint8m1_t vlmul_trunc_u8m1(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m8_u8m1))) +vuint8m1_t vlmul_trunc_u8m1(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m2_u16m1))) +vuint16m1_t vlmul_trunc_u16m1(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m4_u16m1))) +vuint16m1_t vlmul_trunc_u16m1(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m8_u16m1))) +vuint16m1_t vlmul_trunc_u16m1(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m2_u32m1))) +vuint32m1_t vlmul_trunc_u32m1(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m4_u32m1))) +vuint32m1_t vlmul_trunc_u32m1(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m8_u32m1))) +vuint32m1_t vlmul_trunc_u32m1(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u64m2_u64m1))) +vuint64m1_t vlmul_trunc_u64m1(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u64m4_u64m1))) +vuint64m1_t vlmul_trunc_u64m1(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u64m8_u64m1))) +vuint64m1_t vlmul_trunc_u64m1(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m4_u8m2))) +vuint8m2_t vlmul_trunc_u8m2(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m8_u8m2))) +vuint8m2_t vlmul_trunc_u8m2(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m4_u16m2))) +vuint16m2_t vlmul_trunc_u16m2(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m8_u16m2))) +vuint16m2_t vlmul_trunc_u16m2(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m4_u32m2))) +vuint32m2_t vlmul_trunc_u32m2(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m8_u32m2))) +vuint32m2_t vlmul_trunc_u32m2(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u64m4_u64m2))) +vuint64m2_t vlmul_trunc_u64m2(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u64m8_u64m2))) +vuint64m2_t vlmul_trunc_u64m2(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m8_u8m4))) +vuint8m4_t vlmul_trunc_u8m4(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m8_u16m4))) +vuint16m4_t vlmul_trunc_u16m4(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m8_u32m4))) +vuint32m4_t vlmul_trunc_u32m4(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u64m8_u64m4))) +vuint64m4_t vlmul_trunc_u64m4(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b8))) +vbool8_t vmand(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b4))) +vbool4_t vmand(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b2))) +vbool2_t vmand(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b1))) +vbool1_t vmand(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b16))) +vbool16_t vmand(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b32))) +vbool32_t vmand(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b64))) +vbool64_t vmand(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b8))) +vbool8_t vmandnot(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b4))) +vbool4_t vmandnot(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b2))) +vbool2_t vmandnot(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b1))) +vbool1_t vmandnot(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b16))) +vbool16_t vmandnot(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b32))) +vbool32_t vmandnot(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b64))) +vbool64_t vmandnot(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b8))) +vbool8_t vmnand(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b4))) +vbool4_t vmnand(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b2))) +vbool2_t vmnand(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b1))) +vbool1_t vmnand(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b16))) +vbool16_t vmnand(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b32))) +vbool32_t vmnand(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b64))) +vbool64_t vmnand(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b8))) +vbool8_t vmnor(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b4))) +vbool4_t vmnor(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b2))) +vbool2_t vmnor(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b1))) +vbool1_t vmnor(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b16))) +vbool16_t vmnor(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b32))) +vbool32_t vmnor(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b64))) +vbool64_t vmnor(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b8))) +vbool8_t vmor(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b4))) +vbool4_t vmor(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b2))) +vbool2_t vmor(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b1))) +vbool1_t vmor(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b16))) +vbool16_t vmor(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b32))) +vbool32_t vmor(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b64))) +vbool64_t vmor(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b8))) +vbool8_t vmornot(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b4))) +vbool4_t vmornot(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b2))) +vbool2_t vmornot(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b1))) +vbool1_t vmornot(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b16))) +vbool16_t vmornot(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b32))) +vbool32_t vmornot(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b64))) +vbool64_t vmornot(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b8))) +vbool8_t vmsbf(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b8_m))) +vbool8_t vmsbf(vbool8_t op0, vbool8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b4))) +vbool4_t vmsbf(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b4_m))) +vbool4_t vmsbf(vbool4_t op0, vbool4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b2))) +vbool2_t vmsbf(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b2_m))) +vbool2_t vmsbf(vbool2_t op0, vbool2_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b1))) +vbool1_t vmsbf(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b1_m))) +vbool1_t vmsbf(vbool1_t op0, vbool1_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b16))) +vbool16_t vmsbf(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b16_m))) +vbool16_t vmsbf(vbool16_t op0, vbool16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b32))) +vbool32_t vmsbf(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b32_m))) +vbool32_t vmsbf(vbool32_t op0, vbool32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b64))) +vbool64_t vmsbf(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b64_m))) +vbool64_t vmsbf(vbool64_t op0, vbool64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b8))) +vbool8_t vmsif(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b8_m))) +vbool8_t vmsif(vbool8_t op0, vbool8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b4))) +vbool4_t vmsif(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b4_m))) +vbool4_t vmsif(vbool4_t op0, vbool4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b2))) +vbool2_t vmsif(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b2_m))) +vbool2_t vmsif(vbool2_t op0, vbool2_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b1))) +vbool1_t vmsif(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b1_m))) +vbool1_t vmsif(vbool1_t op0, vbool1_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b16))) +vbool16_t vmsif(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b16_m))) +vbool16_t vmsif(vbool16_t op0, vbool16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b32))) +vbool32_t vmsif(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b32_m))) +vbool32_t vmsif(vbool32_t op0, vbool32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b64))) +vbool64_t vmsif(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b64_m))) +vbool64_t vmsif(vbool64_t op0, vbool64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b8))) +vbool8_t vmsof(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b8_m))) +vbool8_t vmsof(vbool8_t op0, vbool8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b4))) +vbool4_t vmsof(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b4_m))) +vbool4_t vmsof(vbool4_t op0, vbool4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b2))) +vbool2_t vmsof(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b2_m))) +vbool2_t vmsof(vbool2_t op0, vbool2_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b1))) +vbool1_t vmsof(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b1_m))) +vbool1_t vmsof(vbool1_t op0, vbool1_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b16))) +vbool16_t vmsof(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b16_m))) +vbool16_t vmsof(vbool16_t op0, vbool16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b32))) +vbool32_t vmsof(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b32_m))) +vbool32_t vmsof(vbool32_t op0, vbool32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b64))) +vbool64_t vmsof(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b64_m))) +vbool64_t vmsof(vbool64_t op0, vbool64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b8))) +vbool8_t vmxnor(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b4))) +vbool4_t vmxnor(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b2))) +vbool2_t vmxnor(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b1))) +vbool1_t vmxnor(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b16))) +vbool16_t vmxnor(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b32))) +vbool32_t vmxnor(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b64))) +vbool64_t vmxnor(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b8))) +vbool8_t vmxor(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b4))) +vbool4_t vmxor(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b2))) +vbool2_t vmxor(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b1))) +vbool1_t vmxor(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b16))) +vbool16_t vmxor(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b32))) +vbool32_t vmxor(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b64))) +vbool64_t vmxor(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b8))) +unsigned long vpopc(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b8_m))) +unsigned long vpopc(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b4))) +unsigned long vpopc(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b4_m))) +unsigned long vpopc(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b2))) +unsigned long vpopc(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b2_m))) +unsigned long vpopc(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b1))) +unsigned long vpopc(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b1_m))) +unsigned long vpopc(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b16))) +unsigned long vpopc(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b16_m))) +unsigned long vpopc(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b32))) +unsigned long vpopc(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b32_m))) +unsigned long vpopc(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b64))) +unsigned long vpopc(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b64_m))) +unsigned long vpopc(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m1_i16m1))) +vint16m1_t vreinterpret_i16m1(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m2_i16m2))) +vint16m2_t vreinterpret_i16m2(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m4_i16m4))) +vint16m4_t vreinterpret_i16m4(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m8_i16m8))) +vint16m8_t vreinterpret_i16m8(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8mf2_i16mf2))) +vint16mf2_t vreinterpret_i16mf2(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8mf4_i16mf4))) +vint16mf4_t vreinterpret_i16mf4(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m1_i16m1))) +vint16m1_t vreinterpret_i16m1(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m2_i16m2))) +vint16m2_t vreinterpret_i16m2(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m4_i16m4))) +vint16m4_t vreinterpret_i16m4(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m8_i16m8))) +vint16m8_t vreinterpret_i16m8(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32mf2_i16mf2))) +vint16mf2_t vreinterpret_i16mf2(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m1_i16m1))) +vint16m1_t vreinterpret_i16m1(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m2_i16m2))) +vint16m2_t vreinterpret_i16m2(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m4_i16m4))) +vint16m4_t vreinterpret_i16m4(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m8_i16m8))) +vint16m8_t vreinterpret_i16m8(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m1_i32m1))) +vint32m1_t vreinterpret_i32m1(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m2_i32m2))) +vint32m2_t vreinterpret_i32m2(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m4_i32m4))) +vint32m4_t vreinterpret_i32m4(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m8_i32m8))) +vint32m8_t vreinterpret_i32m8(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8mf2_i32mf2))) +vint32mf2_t vreinterpret_i32mf2(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m1_i32m1))) +vint32m1_t vreinterpret_i32m1(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m2_i32m2))) +vint32m2_t vreinterpret_i32m2(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m4_i32m4))) +vint32m4_t vreinterpret_i32m4(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m8_i32m8))) +vint32m8_t vreinterpret_i32m8(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf2_i32mf2))) +vint32mf2_t vreinterpret_i32mf2(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m1_i32m1))) +vint32m1_t vreinterpret_i32m1(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m2_i32m2))) +vint32m2_t vreinterpret_i32m2(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m4_i32m4))) +vint32m4_t vreinterpret_i32m4(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m8_i32m8))) +vint32m8_t vreinterpret_i32m8(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m1_i64m1))) +vint64m1_t vreinterpret_i64m1(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m2_i64m2))) +vint64m2_t vreinterpret_i64m2(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m4_i64m4))) +vint64m4_t vreinterpret_i64m4(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m8_i64m8))) +vint64m8_t vreinterpret_i64m8(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m1_i64m1))) +vint64m1_t vreinterpret_i64m1(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m2_i64m2))) +vint64m2_t vreinterpret_i64m2(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m4_i64m4))) +vint64m4_t vreinterpret_i64m4(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m8_i64m8))) +vint64m8_t vreinterpret_i64m8(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m1_i64m1))) +vint64m1_t vreinterpret_i64m1(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m2_i64m2))) +vint64m2_t vreinterpret_i64m2(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m4_i64m4))) +vint64m4_t vreinterpret_i64m4(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m8_i64m8))) +vint64m8_t vreinterpret_i64m8(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m1_i8m1))) +vint8m1_t vreinterpret_i8m1(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m2_i8m2))) +vint8m2_t vreinterpret_i8m2(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m4_i8m4))) +vint8m4_t vreinterpret_i8m4(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m8_i8m8))) +vint8m8_t vreinterpret_i8m8(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf2_i8mf2))) +vint8mf2_t vreinterpret_i8mf2(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf4_i8mf4))) +vint8mf4_t vreinterpret_i8mf4(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m1_i8m1))) +vint8m1_t vreinterpret_i8m1(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m2_i8m2))) +vint8m2_t vreinterpret_i8m2(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m4_i8m4))) +vint8m4_t vreinterpret_i8m4(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m8_i8m8))) +vint8m8_t vreinterpret_i8m8(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32mf2_i8mf2))) +vint8mf2_t vreinterpret_i8mf2(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m1_i8m1))) +vint8m1_t vreinterpret_i8m1(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m2_i8m2))) +vint8m2_t vreinterpret_i8m2(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m4_i8m4))) +vint8m4_t vreinterpret_i8m4(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m8_i8m8))) +vint8m8_t vreinterpret_i8m8(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m1_i8m1))) +vint8m1_t vreinterpret_i8m1(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m2_i8m2))) +vint8m2_t vreinterpret_i8m2(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m4_i8m4))) +vint8m4_t vreinterpret_i8m4(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m8_i8m8))) +vint8m8_t vreinterpret_i8m8(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8mf2_i8mf2))) +vint8mf2_t vreinterpret_i8mf2(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8mf4_i8mf4))) +vint8mf4_t vreinterpret_i8mf4(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8mf8_i8mf8))) +vint8mf8_t vreinterpret_i8mf8(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m1_i16m1))) +vint16m1_t vreinterpret_i16m1(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m2_i16m2))) +vint16m2_t vreinterpret_i16m2(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m4_i16m4))) +vint16m4_t vreinterpret_i16m4(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m8_i16m8))) +vint16m8_t vreinterpret_i16m8(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf2_i16mf2))) +vint16mf2_t vreinterpret_i16mf2(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf4_i16mf4))) +vint16mf4_t vreinterpret_i16mf4(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m1_i32m1))) +vint32m1_t vreinterpret_i32m1(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m2_i32m2))) +vint32m2_t vreinterpret_i32m2(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m4_i32m4))) +vint32m4_t vreinterpret_i32m4(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m8_i32m8))) +vint32m8_t vreinterpret_i32m8(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32mf2_i32mf2))) +vint32mf2_t vreinterpret_i32mf2(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m1_i64m1))) +vint64m1_t vreinterpret_i64m1(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m2_i64m2))) +vint64m2_t vreinterpret_i64m2(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m4_i64m4))) +vint64m4_t vreinterpret_i64m4(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m8_i64m8))) +vint64m8_t vreinterpret_i64m8(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m1_u16m1))) +vuint16m1_t vreinterpret_u16m1(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m2_u16m2))) +vuint16m2_t vreinterpret_u16m2(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m4_u16m4))) +vuint16m4_t vreinterpret_u16m4(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m8_u16m8))) +vuint16m8_t vreinterpret_u16m8(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8mf2_u16mf2))) +vuint16mf2_t vreinterpret_u16mf2(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8mf4_u16mf4))) +vuint16mf4_t vreinterpret_u16mf4(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m1_u16m1))) +vuint16m1_t vreinterpret_u16m1(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m2_u16m2))) +vuint16m2_t vreinterpret_u16m2(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m4_u16m4))) +vuint16m4_t vreinterpret_u16m4(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m8_u16m8))) +vuint16m8_t vreinterpret_u16m8(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32mf2_u16mf2))) +vuint16mf2_t vreinterpret_u16mf2(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m1_u16m1))) +vuint16m1_t vreinterpret_u16m1(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m2_u16m2))) +vuint16m2_t vreinterpret_u16m2(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m4_u16m4))) +vuint16m4_t vreinterpret_u16m4(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m8_u16m8))) +vuint16m8_t vreinterpret_u16m8(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m1_u32m1))) +vuint32m1_t vreinterpret_u32m1(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m2_u32m2))) +vuint32m2_t vreinterpret_u32m2(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m4_u32m4))) +vuint32m4_t vreinterpret_u32m4(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m8_u32m8))) +vuint32m8_t vreinterpret_u32m8(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8mf2_u32mf2))) +vuint32mf2_t vreinterpret_u32mf2(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m1_u32m1))) +vuint32m1_t vreinterpret_u32m1(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m2_u32m2))) +vuint32m2_t vreinterpret_u32m2(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m4_u32m4))) +vuint32m4_t vreinterpret_u32m4(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m8_u32m8))) +vuint32m8_t vreinterpret_u32m8(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf2_u32mf2))) +vuint32mf2_t vreinterpret_u32mf2(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m1_u32m1))) +vuint32m1_t vreinterpret_u32m1(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m2_u32m2))) +vuint32m2_t vreinterpret_u32m2(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m4_u32m4))) +vuint32m4_t vreinterpret_u32m4(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m8_u32m8))) +vuint32m8_t vreinterpret_u32m8(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m1_u64m1))) +vuint64m1_t vreinterpret_u64m1(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m2_u64m2))) +vuint64m2_t vreinterpret_u64m2(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m4_u64m4))) +vuint64m4_t vreinterpret_u64m4(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m8_u64m8))) +vuint64m8_t vreinterpret_u64m8(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m1_u64m1))) +vuint64m1_t vreinterpret_u64m1(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m2_u64m2))) +vuint64m2_t vreinterpret_u64m2(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m4_u64m4))) +vuint64m4_t vreinterpret_u64m4(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m8_u64m8))) +vuint64m8_t vreinterpret_u64m8(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m1_u64m1))) +vuint64m1_t vreinterpret_u64m1(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m2_u64m2))) +vuint64m2_t vreinterpret_u64m2(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m4_u64m4))) +vuint64m4_t vreinterpret_u64m4(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m8_u64m8))) +vuint64m8_t vreinterpret_u64m8(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m1_u8m1))) +vuint8m1_t vreinterpret_u8m1(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m2_u8m2))) +vuint8m2_t vreinterpret_u8m2(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m4_u8m4))) +vuint8m4_t vreinterpret_u8m4(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m8_u8m8))) +vuint8m8_t vreinterpret_u8m8(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf2_u8mf2))) +vuint8mf2_t vreinterpret_u8mf2(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf4_u8mf4))) +vuint8mf4_t vreinterpret_u8mf4(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m1_u8m1))) +vuint8m1_t vreinterpret_u8m1(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m2_u8m2))) +vuint8m2_t vreinterpret_u8m2(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m4_u8m4))) +vuint8m4_t vreinterpret_u8m4(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m8_u8m8))) +vuint8m8_t vreinterpret_u8m8(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32mf2_u8mf2))) +vuint8mf2_t vreinterpret_u8mf2(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m1_u8m1))) +vuint8m1_t vreinterpret_u8m1(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m2_u8m2))) +vuint8m2_t vreinterpret_u8m2(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m4_u8m4))) +vuint8m4_t vreinterpret_u8m4(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m8_u8m8))) +vuint8m8_t vreinterpret_u8m8(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m1_u8m1))) +vuint8m1_t vreinterpret_u8m1(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m2_u8m2))) +vuint8m2_t vreinterpret_u8m2(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m4_u8m4))) +vuint8m4_t vreinterpret_u8m4(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m8_u8m8))) +vuint8m8_t vreinterpret_u8m8(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8mf2_u8mf2))) +vuint8mf2_t vreinterpret_u8mf2(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8mf4_u8mf4))) +vuint8mf4_t vreinterpret_u8mf4(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8mf8_u8mf8))) +vuint8mf8_t vreinterpret_u8mf8(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m1_u16m1))) +vuint16m1_t vreinterpret_u16m1(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m2_u16m2))) +vuint16m2_t vreinterpret_u16m2(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m4_u16m4))) +vuint16m4_t vreinterpret_u16m4(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m8_u16m8))) +vuint16m8_t vreinterpret_u16m8(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf2_u16mf2))) +vuint16mf2_t vreinterpret_u16mf2(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf4_u16mf4))) +vuint16mf4_t vreinterpret_u16mf4(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m1_u32m1))) +vuint32m1_t vreinterpret_u32m1(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m2_u32m2))) +vuint32m2_t vreinterpret_u32m2(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m4_u32m4))) +vuint32m4_t vreinterpret_u32m4(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m8_u32m8))) +vuint32m8_t vreinterpret_u32m8(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32mf2_u32mf2))) +vuint32mf2_t vreinterpret_u32mf2(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m1_u64m1))) +vuint64m1_t vreinterpret_u64m1(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m2_u64m2))) +vuint64m2_t vreinterpret_u64m2(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m4_u64m4))) +vuint64m4_t vreinterpret_u64m4(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m8_u64m8))) +vuint64m8_t vreinterpret_u64m8(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b8))) +void vse1(uint8_t * op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b4))) +void vse1(uint8_t * op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b2))) +void vse1(uint8_t * op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b1))) +void vse1(uint8_t * op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b16))) +void vse1(uint8_t * op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b32))) +void vse1(uint8_t * op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b64))) +void vse1(uint8_t * op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16mf4))) +vint16mf4_t vsext_vf2(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16mf4_m))) +vint16mf4_t vsext_vf2(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16mf2))) +vint16mf2_t vsext_vf2(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16mf2_m))) +vint16mf2_t vsext_vf2(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m1))) +vint16m1_t vsext_vf2(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m1_m))) +vint16m1_t vsext_vf2(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m2))) +vint16m2_t vsext_vf2(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m2_m))) +vint16m2_t vsext_vf2(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m4))) +vint16m4_t vsext_vf2(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m4_m))) +vint16m4_t vsext_vf2(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m8))) +vint16m8_t vsext_vf2(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m8_m))) +vint16m8_t vsext_vf2(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32mf2))) +vint32mf2_t vsext_vf2(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32mf2_m))) +vint32mf2_t vsext_vf2(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m1))) +vint32m1_t vsext_vf2(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m1_m))) +vint32m1_t vsext_vf2(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m2))) +vint32m2_t vsext_vf2(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m2_m))) +vint32m2_t vsext_vf2(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m4))) +vint32m4_t vsext_vf2(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m4_m))) +vint32m4_t vsext_vf2(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m8))) +vint32m8_t vsext_vf2(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m8_m))) +vint32m8_t vsext_vf2(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m1))) +vint64m1_t vsext_vf2(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m1_m))) +vint64m1_t vsext_vf2(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m2))) +vint64m2_t vsext_vf2(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m2_m))) +vint64m2_t vsext_vf2(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m4))) +vint64m4_t vsext_vf2(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m4_m))) +vint64m4_t vsext_vf2(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m8))) +vint64m8_t vsext_vf2(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m8_m))) +vint64m8_t vsext_vf2(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32mf2))) +vint32mf2_t vsext_vf4(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32mf2_m))) +vint32mf2_t vsext_vf4(vbool64_t op0, vint32mf2_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m1))) +vint32m1_t vsext_vf4(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m1_m))) +vint32m1_t vsext_vf4(vbool32_t op0, vint32m1_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m2))) +vint32m2_t vsext_vf4(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m2_m))) +vint32m2_t vsext_vf4(vbool16_t op0, vint32m2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m4))) +vint32m4_t vsext_vf4(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m4_m))) +vint32m4_t vsext_vf4(vbool8_t op0, vint32m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m8))) +vint32m8_t vsext_vf4(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m8_m))) +vint32m8_t vsext_vf4(vbool4_t op0, vint32m8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m1))) +vint64m1_t vsext_vf4(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m1_m))) +vint64m1_t vsext_vf4(vbool64_t op0, vint64m1_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m2))) +vint64m2_t vsext_vf4(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m2_m))) +vint64m2_t vsext_vf4(vbool32_t op0, vint64m2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m4))) +vint64m4_t vsext_vf4(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m4_m))) +vint64m4_t vsext_vf4(vbool16_t op0, vint64m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m8))) +vint64m8_t vsext_vf4(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m8_m))) +vint64m8_t vsext_vf4(vbool8_t op0, vint64m8_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m1))) +vint64m1_t vsext_vf8(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m1_m))) +vint64m1_t vsext_vf8(vbool64_t op0, vint64m1_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m2))) +vint64m2_t vsext_vf8(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m2_m))) +vint64m2_t vsext_vf8(vbool32_t op0, vint64m2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m4))) +vint64m4_t vsext_vf8(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m4_m))) +vint64m4_t vsext_vf8(vbool16_t op0, vint64m4_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m8))) +vint64m8_t vsext_vf8(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m8_m))) +vint64m8_t vsext_vf8(vbool8_t op0, vint64m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16mf4))) +vuint16mf4_t vzext_vf2(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16mf4_m))) +vuint16mf4_t vzext_vf2(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16mf2))) +vuint16mf2_t vzext_vf2(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16mf2_m))) +vuint16mf2_t vzext_vf2(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m1))) +vuint16m1_t vzext_vf2(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m1_m))) +vuint16m1_t vzext_vf2(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m2))) +vuint16m2_t vzext_vf2(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m2_m))) +vuint16m2_t vzext_vf2(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m4))) +vuint16m4_t vzext_vf2(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m4_m))) +vuint16m4_t vzext_vf2(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m8))) +vuint16m8_t vzext_vf2(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m8_m))) +vuint16m8_t vzext_vf2(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32mf2))) +vuint32mf2_t vzext_vf2(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32mf2_m))) +vuint32mf2_t vzext_vf2(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m1))) +vuint32m1_t vzext_vf2(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m1_m))) +vuint32m1_t vzext_vf2(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m2))) +vuint32m2_t vzext_vf2(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m2_m))) +vuint32m2_t vzext_vf2(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m4))) +vuint32m4_t vzext_vf2(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m4_m))) +vuint32m4_t vzext_vf2(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m8))) +vuint32m8_t vzext_vf2(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m8_m))) +vuint32m8_t vzext_vf2(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m1))) +vuint64m1_t vzext_vf2(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m1_m))) +vuint64m1_t vzext_vf2(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m2))) +vuint64m2_t vzext_vf2(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m2_m))) +vuint64m2_t vzext_vf2(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m4))) +vuint64m4_t vzext_vf2(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m4_m))) +vuint64m4_t vzext_vf2(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m8))) +vuint64m8_t vzext_vf2(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m8_m))) +vuint64m8_t vzext_vf2(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32mf2))) +vuint32mf2_t vzext_vf4(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32mf2_m))) +vuint32mf2_t vzext_vf4(vbool64_t op0, vuint32mf2_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m1))) +vuint32m1_t vzext_vf4(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m1_m))) +vuint32m1_t vzext_vf4(vbool32_t op0, vuint32m1_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m2))) +vuint32m2_t vzext_vf4(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m2_m))) +vuint32m2_t vzext_vf4(vbool16_t op0, vuint32m2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m4))) +vuint32m4_t vzext_vf4(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m4_m))) +vuint32m4_t vzext_vf4(vbool8_t op0, vuint32m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m8))) +vuint32m8_t vzext_vf4(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m8_m))) +vuint32m8_t vzext_vf4(vbool4_t op0, vuint32m8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m1))) +vuint64m1_t vzext_vf4(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m1_m))) +vuint64m1_t vzext_vf4(vbool64_t op0, vuint64m1_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m2))) +vuint64m2_t vzext_vf4(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m2_m))) +vuint64m2_t vzext_vf4(vbool32_t op0, vuint64m2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m4))) +vuint64m4_t vzext_vf4(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m4_m))) +vuint64m4_t vzext_vf4(vbool16_t op0, vuint64m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m8))) +vuint64m8_t vzext_vf4(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m8_m))) +vuint64m8_t vzext_vf4(vbool8_t op0, vuint64m8_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m1))) +vuint64m1_t vzext_vf8(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m1_m))) +vuint64m1_t vzext_vf8(vbool64_t op0, vuint64m1_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m2))) +vuint64m2_t vzext_vf8(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m2_m))) +vuint64m2_t vzext_vf8(vbool32_t op0, vuint64m2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m4))) +vuint64m4_t vzext_vf8(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m4_m))) +vuint64m4_t vzext_vf8(vbool16_t op0, vuint64m4_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m8))) +vuint64m8_t vzext_vf8(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m8_m))) +vuint64m8_t vzext_vf8(vbool8_t op0, vuint64m8_t op1, vuint8m1_t op2, size_t op3); + +#if defined(__riscv_f) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m1))) +void vsse32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m1_m))) +void vsse32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m2))) +void vsse32(float * op0, ptrdiff_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m2_m))) +void vsse32(vbool16_t op0, float * op1, ptrdiff_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m4))) +void vsse32(float * op0, ptrdiff_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m4_m))) +void vsse32(vbool8_t op0, float * op1, ptrdiff_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m8))) +void vsse32(float * op0, ptrdiff_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m8_m))) +void vsse32(vbool4_t op0, float * op1, ptrdiff_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32mf2))) +void vsse32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32mf2_m))) +void vsse32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m1))) +vfloat32m1_t vluxei8(const float * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m1_m))) +vfloat32m1_t vluxei8(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m2))) +vfloat32m2_t vluxei8(const float * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m2_m))) +vfloat32m2_t vluxei8(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m4))) +vfloat32m4_t vluxei8(const float * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m4_m))) +vfloat32m4_t vluxei8(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m8))) +vfloat32m8_t vluxei8(const float * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m8_m))) +vfloat32m8_t vluxei8(vbool4_t op0, vfloat32m8_t op1, const float * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32mf2))) +vfloat32mf2_t vluxei8(const float * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32mf2_m))) +vfloat32mf2_t vluxei8(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m1))) +vfloat32m1_t vluxei16(const float * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m1_m))) +vfloat32m1_t vluxei16(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m2))) +vfloat32m2_t vluxei16(const float * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m2_m))) +vfloat32m2_t vluxei16(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m4))) +vfloat32m4_t vluxei16(const float * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m4_m))) +vfloat32m4_t vluxei16(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m8))) +vfloat32m8_t vluxei16(const float * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m8_m))) +vfloat32m8_t vluxei16(vbool4_t op0, vfloat32m8_t op1, const float * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32mf2))) +vfloat32mf2_t vluxei16(const float * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32mf2_m))) +vfloat32mf2_t vluxei16(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m1))) +vfloat32m1_t vluxei32(const float * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m1_m))) +vfloat32m1_t vluxei32(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m2))) +vfloat32m2_t vluxei32(const float * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m2_m))) +vfloat32m2_t vluxei32(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m4))) +vfloat32m4_t vluxei32(const float * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m4_m))) +vfloat32m4_t vluxei32(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m8))) +vfloat32m8_t vluxei32(const float * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m8_m))) +vfloat32m8_t vluxei32(vbool4_t op0, vfloat32m8_t op1, const float * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32mf2))) +vfloat32mf2_t vluxei32(const float * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32mf2_m))) +vfloat32mf2_t vluxei32(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32m1))) +vfloat32m1_t vluxei64(const float * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32m1_m))) +vfloat32m1_t vluxei64(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32m2))) +vfloat32m2_t vluxei64(const float * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32m2_m))) +vfloat32m2_t vluxei64(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32m4))) +vfloat32m4_t vluxei64(const float * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32m4_m))) +vfloat32m4_t vluxei64(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32mf2))) +vfloat32mf2_t vluxei64(const float * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32mf2_m))) +vfloat32mf2_t vluxei64(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m1))) +vfloat32m1_t vloxei8(const float * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m1_m))) +vfloat32m1_t vloxei8(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m2))) +vfloat32m2_t vloxei8(const float * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m2_m))) +vfloat32m2_t vloxei8(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m4))) +vfloat32m4_t vloxei8(const float * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m4_m))) +vfloat32m4_t vloxei8(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m8))) +vfloat32m8_t vloxei8(const float * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m8_m))) +vfloat32m8_t vloxei8(vbool4_t op0, vfloat32m8_t op1, const float * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32mf2))) +vfloat32mf2_t vloxei8(const float * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32mf2_m))) +vfloat32mf2_t vloxei8(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m1))) +vfloat32m1_t vloxei16(const float * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m1_m))) +vfloat32m1_t vloxei16(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m2))) +vfloat32m2_t vloxei16(const float * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m2_m))) +vfloat32m2_t vloxei16(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m4))) +vfloat32m4_t vloxei16(const float * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m4_m))) +vfloat32m4_t vloxei16(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m8))) +vfloat32m8_t vloxei16(const float * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m8_m))) +vfloat32m8_t vloxei16(vbool4_t op0, vfloat32m8_t op1, const float * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32mf2))) +vfloat32mf2_t vloxei16(const float * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32mf2_m))) +vfloat32mf2_t vloxei16(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m1))) +vfloat32m1_t vloxei32(const float * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m1_m))) +vfloat32m1_t vloxei32(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m2))) +vfloat32m2_t vloxei32(const float * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m2_m))) +vfloat32m2_t vloxei32(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m4))) +vfloat32m4_t vloxei32(const float * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m4_m))) +vfloat32m4_t vloxei32(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m8))) +vfloat32m8_t vloxei32(const float * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m8_m))) +vfloat32m8_t vloxei32(vbool4_t op0, vfloat32m8_t op1, const float * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32mf2))) +vfloat32mf2_t vloxei32(const float * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32mf2_m))) +vfloat32mf2_t vloxei32(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32m1))) +vfloat32m1_t vloxei64(const float * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32m1_m))) +vfloat32m1_t vloxei64(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32m2))) +vfloat32m2_t vloxei64(const float * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32m2_m))) +vfloat32m2_t vloxei64(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32m4))) +vfloat32m4_t vloxei64(const float * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32m4_m))) +vfloat32m4_t vloxei64(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32mf2))) +vfloat32mf2_t vloxei64(const float * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32mf2_m))) +vfloat32mf2_t vloxei64(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f32m1))) +vfloat32m1_t vmv_v(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f32m2))) +vfloat32m2_t vmv_v(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f32m4))) +vfloat32m4_t vmv_v(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f32m8))) +vfloat32m8_t vmv_v(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f32mf2))) +vfloat32mf2_t vmv_v(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m1))) +vfloat32m1_t vfadd(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m1_m))) +vfloat32m1_t vfadd(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m2))) +vfloat32m2_t vfadd(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m2_m))) +vfloat32m2_t vfadd(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m4))) +vfloat32m4_t vfadd(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m4_m))) +vfloat32m4_t vfadd(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m8))) +vfloat32m8_t vfadd(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m8_m))) +vfloat32m8_t vfadd(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32mf2))) +vfloat32mf2_t vfadd(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32mf2_m))) +vfloat32mf2_t vfadd(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m1))) +vfloat32m1_t vfadd(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m1_m))) +vfloat32m1_t vfadd(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m2))) +vfloat32m2_t vfadd(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m2_m))) +vfloat32m2_t vfadd(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m4))) +vfloat32m4_t vfadd(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m4_m))) +vfloat32m4_t vfadd(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m8))) +vfloat32m8_t vfadd(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m8_m))) +vfloat32m8_t vfadd(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32mf2))) +vfloat32mf2_t vfadd(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32mf2_m))) +vfloat32mf2_t vfadd(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m1))) +vfloat32m1_t vfsub(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m1_m))) +vfloat32m1_t vfsub(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m2))) +vfloat32m2_t vfsub(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m2_m))) +vfloat32m2_t vfsub(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m4))) +vfloat32m4_t vfsub(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m4_m))) +vfloat32m4_t vfsub(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m8))) +vfloat32m8_t vfsub(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m8_m))) +vfloat32m8_t vfsub(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32mf2))) +vfloat32mf2_t vfsub(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32mf2_m))) +vfloat32mf2_t vfsub(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m1))) +vfloat32m1_t vfsub(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m1_m))) +vfloat32m1_t vfsub(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m2))) +vfloat32m2_t vfsub(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m2_m))) +vfloat32m2_t vfsub(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m4))) +vfloat32m4_t vfsub(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m4_m))) +vfloat32m4_t vfsub(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m8))) +vfloat32m8_t vfsub(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m8_m))) +vfloat32m8_t vfsub(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32mf2))) +vfloat32mf2_t vfsub(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32mf2_m))) +vfloat32mf2_t vfsub(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m1))) +vfloat32m1_t vfrsub(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m1_m))) +vfloat32m1_t vfrsub(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m2))) +vfloat32m2_t vfrsub(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m2_m))) +vfloat32m2_t vfrsub(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m4))) +vfloat32m4_t vfrsub(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m4_m))) +vfloat32m4_t vfrsub(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m8))) +vfloat32m8_t vfrsub(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m8_m))) +vfloat32m8_t vfrsub(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32mf2))) +vfloat32mf2_t vfrsub(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32mf2_m))) +vfloat32mf2_t vfrsub(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m1))) +vfloat32m1_t vfmul(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m1_m))) +vfloat32m1_t vfmul(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m2))) +vfloat32m2_t vfmul(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m2_m))) +vfloat32m2_t vfmul(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m4))) +vfloat32m4_t vfmul(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m4_m))) +vfloat32m4_t vfmul(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m8))) +vfloat32m8_t vfmul(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m8_m))) +vfloat32m8_t vfmul(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32mf2))) +vfloat32mf2_t vfmul(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32mf2_m))) +vfloat32mf2_t vfmul(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m1))) +vfloat32m1_t vfmul(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m1_m))) +vfloat32m1_t vfmul(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m2))) +vfloat32m2_t vfmul(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m2_m))) +vfloat32m2_t vfmul(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m4))) +vfloat32m4_t vfmul(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m4_m))) +vfloat32m4_t vfmul(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m8))) +vfloat32m8_t vfmul(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m8_m))) +vfloat32m8_t vfmul(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32mf2))) +vfloat32mf2_t vfmul(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32mf2_m))) +vfloat32mf2_t vfmul(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m1))) +vfloat32m1_t vfdiv(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m1_m))) +vfloat32m1_t vfdiv(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m2))) +vfloat32m2_t vfdiv(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m2_m))) +vfloat32m2_t vfdiv(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m4))) +vfloat32m4_t vfdiv(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m4_m))) +vfloat32m4_t vfdiv(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m8))) +vfloat32m8_t vfdiv(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m8_m))) +vfloat32m8_t vfdiv(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32mf2))) +vfloat32mf2_t vfdiv(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32mf2_m))) +vfloat32mf2_t vfdiv(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m1))) +vfloat32m1_t vfdiv(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m1_m))) +vfloat32m1_t vfdiv(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m2))) +vfloat32m2_t vfdiv(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m2_m))) +vfloat32m2_t vfdiv(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m4))) +vfloat32m4_t vfdiv(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m4_m))) +vfloat32m4_t vfdiv(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m8))) +vfloat32m8_t vfdiv(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m8_m))) +vfloat32m8_t vfdiv(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32mf2))) +vfloat32mf2_t vfdiv(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32mf2_m))) +vfloat32mf2_t vfdiv(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m1))) +vfloat32m1_t vfrdiv(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m1_m))) +vfloat32m1_t vfrdiv(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m2))) +vfloat32m2_t vfrdiv(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m2_m))) +vfloat32m2_t vfrdiv(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m4))) +vfloat32m4_t vfrdiv(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m4_m))) +vfloat32m4_t vfrdiv(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m8))) +vfloat32m8_t vfrdiv(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m8_m))) +vfloat32m8_t vfrdiv(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32mf2))) +vfloat32mf2_t vfrdiv(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32mf2_m))) +vfloat32mf2_t vfrdiv(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m1))) +vfloat32m1_t vfmacc(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m1_m))) +vfloat32m1_t vfmacc(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m2))) +vfloat32m2_t vfmacc(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m2_m))) +vfloat32m2_t vfmacc(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m4))) +vfloat32m4_t vfmacc(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m4_m))) +vfloat32m4_t vfmacc(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m8))) +vfloat32m8_t vfmacc(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m8_m))) +vfloat32m8_t vfmacc(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32mf2))) +vfloat32mf2_t vfmacc(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32mf2_m))) +vfloat32mf2_t vfmacc(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m1))) +vfloat32m1_t vfmacc(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m1_m))) +vfloat32m1_t vfmacc(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m2))) +vfloat32m2_t vfmacc(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m2_m))) +vfloat32m2_t vfmacc(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m4))) +vfloat32m4_t vfmacc(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m4_m))) +vfloat32m4_t vfmacc(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m8))) +vfloat32m8_t vfmacc(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m8_m))) +vfloat32m8_t vfmacc(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32mf2))) +vfloat32mf2_t vfmacc(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32mf2_m))) +vfloat32mf2_t vfmacc(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m1))) +vfloat32m1_t vfnmacc(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m1_m))) +vfloat32m1_t vfnmacc(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m2))) +vfloat32m2_t vfnmacc(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m2_m))) +vfloat32m2_t vfnmacc(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m4))) +vfloat32m4_t vfnmacc(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m4_m))) +vfloat32m4_t vfnmacc(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m8))) +vfloat32m8_t vfnmacc(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m8_m))) +vfloat32m8_t vfnmacc(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32mf2))) +vfloat32mf2_t vfnmacc(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32mf2_m))) +vfloat32mf2_t vfnmacc(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m1))) +vfloat32m1_t vfnmacc(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m1_m))) +vfloat32m1_t vfnmacc(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m2))) +vfloat32m2_t vfnmacc(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m2_m))) +vfloat32m2_t vfnmacc(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m4))) +vfloat32m4_t vfnmacc(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m4_m))) +vfloat32m4_t vfnmacc(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m8))) +vfloat32m8_t vfnmacc(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m8_m))) +vfloat32m8_t vfnmacc(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32mf2))) +vfloat32mf2_t vfnmacc(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32mf2_m))) +vfloat32mf2_t vfnmacc(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m1))) +vfloat32m1_t vfmsac(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m1_m))) +vfloat32m1_t vfmsac(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m2))) +vfloat32m2_t vfmsac(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m2_m))) +vfloat32m2_t vfmsac(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m4))) +vfloat32m4_t vfmsac(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m4_m))) +vfloat32m4_t vfmsac(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m8))) +vfloat32m8_t vfmsac(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m8_m))) +vfloat32m8_t vfmsac(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32mf2))) +vfloat32mf2_t vfmsac(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32mf2_m))) +vfloat32mf2_t vfmsac(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m1))) +vfloat32m1_t vfmsac(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m1_m))) +vfloat32m1_t vfmsac(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m2))) +vfloat32m2_t vfmsac(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m2_m))) +vfloat32m2_t vfmsac(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m4))) +vfloat32m4_t vfmsac(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m4_m))) +vfloat32m4_t vfmsac(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m8))) +vfloat32m8_t vfmsac(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m8_m))) +vfloat32m8_t vfmsac(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32mf2))) +vfloat32mf2_t vfmsac(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32mf2_m))) +vfloat32mf2_t vfmsac(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m1))) +vfloat32m1_t vfnmsac(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m1_m))) +vfloat32m1_t vfnmsac(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m2))) +vfloat32m2_t vfnmsac(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m2_m))) +vfloat32m2_t vfnmsac(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m4))) +vfloat32m4_t vfnmsac(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m4_m))) +vfloat32m4_t vfnmsac(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m8))) +vfloat32m8_t vfnmsac(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m8_m))) +vfloat32m8_t vfnmsac(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32mf2))) +vfloat32mf2_t vfnmsac(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32mf2_m))) +vfloat32mf2_t vfnmsac(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m1))) +vfloat32m1_t vfnmsac(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m1_m))) +vfloat32m1_t vfnmsac(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m2))) +vfloat32m2_t vfnmsac(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m2_m))) +vfloat32m2_t vfnmsac(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m4))) +vfloat32m4_t vfnmsac(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m4_m))) +vfloat32m4_t vfnmsac(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m8))) +vfloat32m8_t vfnmsac(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m8_m))) +vfloat32m8_t vfnmsac(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32mf2))) +vfloat32mf2_t vfnmsac(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32mf2_m))) +vfloat32mf2_t vfnmsac(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m1))) +vfloat32m1_t vfmadd(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m1_m))) +vfloat32m1_t vfmadd(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m2))) +vfloat32m2_t vfmadd(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m2_m))) +vfloat32m2_t vfmadd(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m4))) +vfloat32m4_t vfmadd(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m4_m))) +vfloat32m4_t vfmadd(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m8))) +vfloat32m8_t vfmadd(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m8_m))) +vfloat32m8_t vfmadd(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32mf2))) +vfloat32mf2_t vfmadd(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32mf2_m))) +vfloat32mf2_t vfmadd(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m1))) +vfloat32m1_t vfmadd(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m1_m))) +vfloat32m1_t vfmadd(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m2))) +vfloat32m2_t vfmadd(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m2_m))) +vfloat32m2_t vfmadd(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m4))) +vfloat32m4_t vfmadd(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m4_m))) +vfloat32m4_t vfmadd(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m8))) +vfloat32m8_t vfmadd(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m8_m))) +vfloat32m8_t vfmadd(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32mf2))) +vfloat32mf2_t vfmadd(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32mf2_m))) +vfloat32mf2_t vfmadd(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m1))) +vfloat32m1_t vfnmadd(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m1_m))) +vfloat32m1_t vfnmadd(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m2))) +vfloat32m2_t vfnmadd(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m2_m))) +vfloat32m2_t vfnmadd(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m4))) +vfloat32m4_t vfnmadd(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m4_m))) +vfloat32m4_t vfnmadd(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m8))) +vfloat32m8_t vfnmadd(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m8_m))) +vfloat32m8_t vfnmadd(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32mf2))) +vfloat32mf2_t vfnmadd(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32mf2_m))) +vfloat32mf2_t vfnmadd(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m1))) +vfloat32m1_t vfnmadd(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m1_m))) +vfloat32m1_t vfnmadd(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m2))) +vfloat32m2_t vfnmadd(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m2_m))) +vfloat32m2_t vfnmadd(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m4))) +vfloat32m4_t vfnmadd(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m4_m))) +vfloat32m4_t vfnmadd(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m8))) +vfloat32m8_t vfnmadd(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m8_m))) +vfloat32m8_t vfnmadd(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32mf2))) +vfloat32mf2_t vfnmadd(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32mf2_m))) +vfloat32mf2_t vfnmadd(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m1))) +vfloat32m1_t vfmsub(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m1_m))) +vfloat32m1_t vfmsub(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m2))) +vfloat32m2_t vfmsub(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m2_m))) +vfloat32m2_t vfmsub(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m4))) +vfloat32m4_t vfmsub(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m4_m))) +vfloat32m4_t vfmsub(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m8))) +vfloat32m8_t vfmsub(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m8_m))) +vfloat32m8_t vfmsub(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32mf2))) +vfloat32mf2_t vfmsub(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32mf2_m))) +vfloat32mf2_t vfmsub(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m1))) +vfloat32m1_t vfmsub(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m1_m))) +vfloat32m1_t vfmsub(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m2))) +vfloat32m2_t vfmsub(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m2_m))) +vfloat32m2_t vfmsub(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m4))) +vfloat32m4_t vfmsub(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m4_m))) +vfloat32m4_t vfmsub(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m8))) +vfloat32m8_t vfmsub(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m8_m))) +vfloat32m8_t vfmsub(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32mf2))) +vfloat32mf2_t vfmsub(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32mf2_m))) +vfloat32mf2_t vfmsub(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m1))) +vfloat32m1_t vfnmsub(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m1_m))) +vfloat32m1_t vfnmsub(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m2))) +vfloat32m2_t vfnmsub(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m2_m))) +vfloat32m2_t vfnmsub(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m4))) +vfloat32m4_t vfnmsub(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m4_m))) +vfloat32m4_t vfnmsub(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m8))) +vfloat32m8_t vfnmsub(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m8_m))) +vfloat32m8_t vfnmsub(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32mf2))) +vfloat32mf2_t vfnmsub(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32mf2_m))) +vfloat32mf2_t vfnmsub(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m1))) +vfloat32m1_t vfnmsub(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m1_m))) +vfloat32m1_t vfnmsub(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m2))) +vfloat32m2_t vfnmsub(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m2_m))) +vfloat32m2_t vfnmsub(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m4))) +vfloat32m4_t vfnmsub(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m4_m))) +vfloat32m4_t vfnmsub(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m8))) +vfloat32m8_t vfnmsub(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m8_m))) +vfloat32m8_t vfnmsub(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32mf2))) +vfloat32mf2_t vfnmsub(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32mf2_m))) +vfloat32mf2_t vfnmsub(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m1))) +vfloat32m1_t vfmin(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m1_m))) +vfloat32m1_t vfmin(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m2))) +vfloat32m2_t vfmin(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m2_m))) +vfloat32m2_t vfmin(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m4))) +vfloat32m4_t vfmin(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m4_m))) +vfloat32m4_t vfmin(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m8))) +vfloat32m8_t vfmin(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m8_m))) +vfloat32m8_t vfmin(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32mf2))) +vfloat32mf2_t vfmin(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32mf2_m))) +vfloat32mf2_t vfmin(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m1))) +vfloat32m1_t vfmin(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m1_m))) +vfloat32m1_t vfmin(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m2))) +vfloat32m2_t vfmin(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m2_m))) +vfloat32m2_t vfmin(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m4))) +vfloat32m4_t vfmin(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m4_m))) +vfloat32m4_t vfmin(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m8))) +vfloat32m8_t vfmin(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m8_m))) +vfloat32m8_t vfmin(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32mf2))) +vfloat32mf2_t vfmin(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32mf2_m))) +vfloat32mf2_t vfmin(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m1))) +vfloat32m1_t vfmax(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m1_m))) +vfloat32m1_t vfmax(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m2))) +vfloat32m2_t vfmax(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m2_m))) +vfloat32m2_t vfmax(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m4))) +vfloat32m4_t vfmax(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m4_m))) +vfloat32m4_t vfmax(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m8))) +vfloat32m8_t vfmax(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m8_m))) +vfloat32m8_t vfmax(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32mf2))) +vfloat32mf2_t vfmax(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32mf2_m))) +vfloat32mf2_t vfmax(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m1))) +vfloat32m1_t vfmax(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m1_m))) +vfloat32m1_t vfmax(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m2))) +vfloat32m2_t vfmax(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m2_m))) +vfloat32m2_t vfmax(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m4))) +vfloat32m4_t vfmax(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m4_m))) +vfloat32m4_t vfmax(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m8))) +vfloat32m8_t vfmax(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m8_m))) +vfloat32m8_t vfmax(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32mf2))) +vfloat32mf2_t vfmax(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32mf2_m))) +vfloat32mf2_t vfmax(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m1))) +vfloat32m1_t vfsgnj(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m1_m))) +vfloat32m1_t vfsgnj(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m2))) +vfloat32m2_t vfsgnj(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m2_m))) +vfloat32m2_t vfsgnj(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m4))) +vfloat32m4_t vfsgnj(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m4_m))) +vfloat32m4_t vfsgnj(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m8))) +vfloat32m8_t vfsgnj(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m8_m))) +vfloat32m8_t vfsgnj(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32mf2))) +vfloat32mf2_t vfsgnj(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32mf2_m))) +vfloat32mf2_t vfsgnj(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m1))) +vfloat32m1_t vfsgnj(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m1_m))) +vfloat32m1_t vfsgnj(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m2))) +vfloat32m2_t vfsgnj(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m2_m))) +vfloat32m2_t vfsgnj(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m4))) +vfloat32m4_t vfsgnj(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m4_m))) +vfloat32m4_t vfsgnj(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m8))) +vfloat32m8_t vfsgnj(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m8_m))) +vfloat32m8_t vfsgnj(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32mf2))) +vfloat32mf2_t vfsgnj(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32mf2_m))) +vfloat32mf2_t vfsgnj(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m1))) +vfloat32m1_t vfsgnjn(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m1_m))) +vfloat32m1_t vfsgnjn(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m2))) +vfloat32m2_t vfsgnjn(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m2_m))) +vfloat32m2_t vfsgnjn(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m4))) +vfloat32m4_t vfsgnjn(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m4_m))) +vfloat32m4_t vfsgnjn(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m8))) +vfloat32m8_t vfsgnjn(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m8_m))) +vfloat32m8_t vfsgnjn(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32mf2))) +vfloat32mf2_t vfsgnjn(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32mf2_m))) +vfloat32mf2_t vfsgnjn(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m1))) +vfloat32m1_t vfsgnjn(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m1_m))) +vfloat32m1_t vfsgnjn(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m2))) +vfloat32m2_t vfsgnjn(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m2_m))) +vfloat32m2_t vfsgnjn(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m4))) +vfloat32m4_t vfsgnjn(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m4_m))) +vfloat32m4_t vfsgnjn(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m8))) +vfloat32m8_t vfsgnjn(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m8_m))) +vfloat32m8_t vfsgnjn(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32mf2))) +vfloat32mf2_t vfsgnjn(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32mf2_m))) +vfloat32mf2_t vfsgnjn(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m1))) +vfloat32m1_t vfsgnjx(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m1_m))) +vfloat32m1_t vfsgnjx(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m2))) +vfloat32m2_t vfsgnjx(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m2_m))) +vfloat32m2_t vfsgnjx(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m4))) +vfloat32m4_t vfsgnjx(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m4_m))) +vfloat32m4_t vfsgnjx(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m8))) +vfloat32m8_t vfsgnjx(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m8_m))) +vfloat32m8_t vfsgnjx(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32mf2))) +vfloat32mf2_t vfsgnjx(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32mf2_m))) +vfloat32mf2_t vfsgnjx(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m1))) +vfloat32m1_t vfsgnjx(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m1_m))) +vfloat32m1_t vfsgnjx(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m2))) +vfloat32m2_t vfsgnjx(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m2_m))) +vfloat32m2_t vfsgnjx(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m4))) +vfloat32m4_t vfsgnjx(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m4_m))) +vfloat32m4_t vfsgnjx(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m8))) +vfloat32m8_t vfsgnjx(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m8_m))) +vfloat32m8_t vfsgnjx(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32mf2))) +vfloat32mf2_t vfsgnjx(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32mf2_m))) +vfloat32mf2_t vfsgnjx(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m1))) +vfloat32m1_t vfabs(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m1_m))) +vfloat32m1_t vfabs(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m2))) +vfloat32m2_t vfabs(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m2_m))) +vfloat32m2_t vfabs(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m4))) +vfloat32m4_t vfabs(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m4_m))) +vfloat32m4_t vfabs(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m8))) +vfloat32m8_t vfabs(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m8_m))) +vfloat32m8_t vfabs(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32mf2))) +vfloat32mf2_t vfabs(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32mf2_m))) +vfloat32mf2_t vfabs(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m1_b32))) +vbool32_t vmfeq(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m1_b32_m))) +vbool32_t vmfeq(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m2_b16))) +vbool16_t vmfeq(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m2_b16_m))) +vbool16_t vmfeq(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m4_b8))) +vbool8_t vmfeq(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m4_b8_m))) +vbool8_t vmfeq(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m8_b4))) +vbool4_t vmfeq(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m8_b4_m))) +vbool4_t vmfeq(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32mf2_b64))) +vbool64_t vmfeq(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32mf2_b64_m))) +vbool64_t vmfeq(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m1_b32))) +vbool32_t vmfeq(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m1_b32_m))) +vbool32_t vmfeq(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m2_b16))) +vbool16_t vmfeq(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m2_b16_m))) +vbool16_t vmfeq(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m4_b8))) +vbool8_t vmfeq(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m4_b8_m))) +vbool8_t vmfeq(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m8_b4))) +vbool4_t vmfeq(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m8_b4_m))) +vbool4_t vmfeq(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32mf2_b64))) +vbool64_t vmfeq(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32mf2_b64_m))) +vbool64_t vmfeq(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m1_b32))) +vbool32_t vmfne(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m1_b32_m))) +vbool32_t vmfne(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m2_b16))) +vbool16_t vmfne(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m2_b16_m))) +vbool16_t vmfne(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m4_b8))) +vbool8_t vmfne(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m4_b8_m))) +vbool8_t vmfne(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m8_b4))) +vbool4_t vmfne(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m8_b4_m))) +vbool4_t vmfne(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32mf2_b64))) +vbool64_t vmfne(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32mf2_b64_m))) +vbool64_t vmfne(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m1_b32))) +vbool32_t vmfne(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m1_b32_m))) +vbool32_t vmfne(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m2_b16))) +vbool16_t vmfne(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m2_b16_m))) +vbool16_t vmfne(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m4_b8))) +vbool8_t vmfne(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m4_b8_m))) +vbool8_t vmfne(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m8_b4))) +vbool4_t vmfne(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m8_b4_m))) +vbool4_t vmfne(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32mf2_b64))) +vbool64_t vmfne(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32mf2_b64_m))) +vbool64_t vmfne(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m1_b32))) +vbool32_t vmflt(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m1_b32_m))) +vbool32_t vmflt(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m2_b16))) +vbool16_t vmflt(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m2_b16_m))) +vbool16_t vmflt(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m4_b8))) +vbool8_t vmflt(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m4_b8_m))) +vbool8_t vmflt(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m8_b4))) +vbool4_t vmflt(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m8_b4_m))) +vbool4_t vmflt(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32mf2_b64))) +vbool64_t vmflt(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32mf2_b64_m))) +vbool64_t vmflt(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m1_b32))) +vbool32_t vmflt(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m1_b32_m))) +vbool32_t vmflt(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m2_b16))) +vbool16_t vmflt(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m2_b16_m))) +vbool16_t vmflt(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m4_b8))) +vbool8_t vmflt(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m4_b8_m))) +vbool8_t vmflt(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m8_b4))) +vbool4_t vmflt(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m8_b4_m))) +vbool4_t vmflt(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32mf2_b64))) +vbool64_t vmflt(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32mf2_b64_m))) +vbool64_t vmflt(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m1_b32))) +vbool32_t vmfle(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m1_b32_m))) +vbool32_t vmfle(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m2_b16))) +vbool16_t vmfle(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m2_b16_m))) +vbool16_t vmfle(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m4_b8))) +vbool8_t vmfle(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m4_b8_m))) +vbool8_t vmfle(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m8_b4))) +vbool4_t vmfle(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m8_b4_m))) +vbool4_t vmfle(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32mf2_b64))) +vbool64_t vmfle(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32mf2_b64_m))) +vbool64_t vmfle(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m1_b32))) +vbool32_t vmfle(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m1_b32_m))) +vbool32_t vmfle(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m2_b16))) +vbool16_t vmfle(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m2_b16_m))) +vbool16_t vmfle(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m4_b8))) +vbool8_t vmfle(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m4_b8_m))) +vbool8_t vmfle(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m8_b4))) +vbool4_t vmfle(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m8_b4_m))) +vbool4_t vmfle(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32mf2_b64))) +vbool64_t vmfle(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32mf2_b64_m))) +vbool64_t vmfle(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m1_b32))) +vbool32_t vmfgt(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m1_b32_m))) +vbool32_t vmfgt(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m2_b16))) +vbool16_t vmfgt(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m2_b16_m))) +vbool16_t vmfgt(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m4_b8))) +vbool8_t vmfgt(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m4_b8_m))) +vbool8_t vmfgt(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m8_b4))) +vbool4_t vmfgt(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m8_b4_m))) +vbool4_t vmfgt(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32mf2_b64))) +vbool64_t vmfgt(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32mf2_b64_m))) +vbool64_t vmfgt(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m1_b32))) +vbool32_t vmfgt(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m1_b32_m))) +vbool32_t vmfgt(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m2_b16))) +vbool16_t vmfgt(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m2_b16_m))) +vbool16_t vmfgt(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m4_b8))) +vbool8_t vmfgt(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m4_b8_m))) +vbool8_t vmfgt(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m8_b4))) +vbool4_t vmfgt(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m8_b4_m))) +vbool4_t vmfgt(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32mf2_b64))) +vbool64_t vmfgt(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32mf2_b64_m))) +vbool64_t vmfgt(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m1_b32))) +vbool32_t vmfge(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m1_b32_m))) +vbool32_t vmfge(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m2_b16))) +vbool16_t vmfge(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m2_b16_m))) +vbool16_t vmfge(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m4_b8))) +vbool8_t vmfge(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m4_b8_m))) +vbool8_t vmfge(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m8_b4))) +vbool4_t vmfge(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m8_b4_m))) +vbool4_t vmfge(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32mf2_b64))) +vbool64_t vmfge(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32mf2_b64_m))) +vbool64_t vmfge(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m1_b32))) +vbool32_t vmfge(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m1_b32_m))) +vbool32_t vmfge(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m2_b16))) +vbool16_t vmfge(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m2_b16_m))) +vbool16_t vmfge(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m4_b8))) +vbool8_t vmfge(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m4_b8_m))) +vbool8_t vmfge(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m8_b4))) +vbool4_t vmfge(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m8_b4_m))) +vbool4_t vmfge(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32mf2_b64))) +vbool64_t vmfge(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32mf2_b64_m))) +vbool64_t vmfge(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f32m1))) +vfloat32m1_t vmerge(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f32m2))) +vfloat32m2_t vmerge(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f32m4))) +vfloat32m4_t vmerge(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f32m8))) +vfloat32m8_t vmerge(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f32mf2))) +vfloat32mf2_t vmerge(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f32m1))) +vfloat32m1_t vfmerge(vbool32_t op0, vfloat32m1_t op1, float op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f32m2))) +vfloat32m2_t vfmerge(vbool16_t op0, vfloat32m2_t op1, float op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f32m4))) +vfloat32m4_t vfmerge(vbool8_t op0, vfloat32m4_t op1, float op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f32m8))) +vfloat32m8_t vfmerge(vbool4_t op0, vfloat32m8_t op1, float op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f32mf2))) +vfloat32mf2_t vfmerge(vbool64_t op0, vfloat32mf2_t op1, float op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m1_f32m1))) +vfloat32m1_t vfredmax(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m1_f32m1_m))) +vfloat32m1_t vfredmax(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m2_f32m1))) +vfloat32m1_t vfredmax(vfloat32m1_t op0, vfloat32m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m2_f32m1_m))) +vfloat32m1_t vfredmax(vbool16_t op0, vfloat32m1_t op1, vfloat32m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m4_f32m1))) +vfloat32m1_t vfredmax(vfloat32m1_t op0, vfloat32m4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m4_f32m1_m))) +vfloat32m1_t vfredmax(vbool8_t op0, vfloat32m1_t op1, vfloat32m4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m8_f32m1))) +vfloat32m1_t vfredmax(vfloat32m1_t op0, vfloat32m8_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m8_f32m1_m))) +vfloat32m1_t vfredmax(vbool4_t op0, vfloat32m1_t op1, vfloat32m8_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32mf2_f32m1))) +vfloat32m1_t vfredmax(vfloat32m1_t op0, vfloat32mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32mf2_f32m1_m))) +vfloat32m1_t vfredmax(vbool64_t op0, vfloat32m1_t op1, vfloat32mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m1_f32m1))) +vfloat32m1_t vfredmin(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m1_f32m1_m))) +vfloat32m1_t vfredmin(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m2_f32m1))) +vfloat32m1_t vfredmin(vfloat32m1_t op0, vfloat32m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m2_f32m1_m))) +vfloat32m1_t vfredmin(vbool16_t op0, vfloat32m1_t op1, vfloat32m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m4_f32m1))) +vfloat32m1_t vfredmin(vfloat32m1_t op0, vfloat32m4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m4_f32m1_m))) +vfloat32m1_t vfredmin(vbool8_t op0, vfloat32m1_t op1, vfloat32m4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m8_f32m1))) +vfloat32m1_t vfredmin(vfloat32m1_t op0, vfloat32m8_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m8_f32m1_m))) +vfloat32m1_t vfredmin(vbool4_t op0, vfloat32m1_t op1, vfloat32m8_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32mf2_f32m1))) +vfloat32m1_t vfredmin(vfloat32m1_t op0, vfloat32mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32mf2_f32m1_m))) +vfloat32m1_t vfredmin(vbool64_t op0, vfloat32m1_t op1, vfloat32mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m1_f32m1))) +vfloat32m1_t vfredsum(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m1_f32m1_m))) +vfloat32m1_t vfredsum(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m2_f32m1))) +vfloat32m1_t vfredsum(vfloat32m1_t op0, vfloat32m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m2_f32m1_m))) +vfloat32m1_t vfredsum(vbool16_t op0, vfloat32m1_t op1, vfloat32m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m4_f32m1))) +vfloat32m1_t vfredsum(vfloat32m1_t op0, vfloat32m4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m4_f32m1_m))) +vfloat32m1_t vfredsum(vbool8_t op0, vfloat32m1_t op1, vfloat32m4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m8_f32m1))) +vfloat32m1_t vfredsum(vfloat32m1_t op0, vfloat32m8_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m8_f32m1_m))) +vfloat32m1_t vfredsum(vbool4_t op0, vfloat32m1_t op1, vfloat32m8_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32mf2_f32m1))) +vfloat32m1_t vfredsum(vfloat32m1_t op0, vfloat32mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32mf2_f32m1_m))) +vfloat32m1_t vfredsum(vbool64_t op0, vfloat32m1_t op1, vfloat32mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m1_f32m1))) +vfloat32m1_t vfredosum(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m1_f32m1_m))) +vfloat32m1_t vfredosum(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m2_f32m1))) +vfloat32m1_t vfredosum(vfloat32m1_t op0, vfloat32m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m2_f32m1_m))) +vfloat32m1_t vfredosum(vbool16_t op0, vfloat32m1_t op1, vfloat32m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m4_f32m1))) +vfloat32m1_t vfredosum(vfloat32m1_t op0, vfloat32m4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m4_f32m1_m))) +vfloat32m1_t vfredosum(vbool8_t op0, vfloat32m1_t op1, vfloat32m4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m8_f32m1))) +vfloat32m1_t vfredosum(vfloat32m1_t op0, vfloat32m8_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m8_f32m1_m))) +vfloat32m1_t vfredosum(vbool4_t op0, vfloat32m1_t op1, vfloat32m8_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32mf2_f32m1))) +vfloat32m1_t vfredosum(vfloat32m1_t op0, vfloat32mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32mf2_f32m1_m))) +vfloat32m1_t vfredosum(vbool64_t op0, vfloat32m1_t op1, vfloat32mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f32m1_f32))) +float vfmv_f(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f32m2_f32))) +float vfmv_f(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f32m4_f32))) +float vfmv_f(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f32m8_f32))) +float vfmv_f(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f32mf2_f32))) +float vfmv_f(vfloat32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f32m1))) +vfloat32m1_t vfmv_s(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f32m2))) +vfloat32m2_t vfmv_s(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f32m4))) +vfloat32m4_t vfmv_s(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f32m8))) +vfloat32m8_t vfmv_s(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f32mf2))) +vfloat32mf2_t vfmv_s(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m1))) +vfloat32m1_t vslideup(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m1_m))) +vfloat32m1_t vslideup(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m2))) +vfloat32m2_t vslideup(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m2_m))) +vfloat32m2_t vslideup(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m4))) +vfloat32m4_t vslideup(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m4_m))) +vfloat32m4_t vslideup(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m8))) +vfloat32m8_t vslideup(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m8_m))) +vfloat32m8_t vslideup(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32mf2))) +vfloat32mf2_t vslideup(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32mf2_m))) +vfloat32mf2_t vslideup(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m1))) +vfloat32m1_t vslidedown(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m1_m))) +vfloat32m1_t vslidedown(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m2))) +vfloat32m2_t vslidedown(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m2_m))) +vfloat32m2_t vslidedown(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m4))) +vfloat32m4_t vslidedown(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m4_m))) +vfloat32m4_t vslidedown(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m8))) +vfloat32m8_t vslidedown(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m8_m))) +vfloat32m8_t vslidedown(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32mf2))) +vfloat32mf2_t vslidedown(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32mf2_m))) +vfloat32mf2_t vslidedown(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m1))) +vfloat32m1_t vfslide1up(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m1_m))) +vfloat32m1_t vfslide1up(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m2))) +vfloat32m2_t vfslide1up(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m2_m))) +vfloat32m2_t vfslide1up(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m4))) +vfloat32m4_t vfslide1up(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m4_m))) +vfloat32m4_t vfslide1up(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m8))) +vfloat32m8_t vfslide1up(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m8_m))) +vfloat32m8_t vfslide1up(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32mf2))) +vfloat32mf2_t vfslide1up(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32mf2_m))) +vfloat32mf2_t vfslide1up(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m1))) +vfloat32m1_t vfslide1down(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m1_m))) +vfloat32m1_t vfslide1down(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m2))) +vfloat32m2_t vfslide1down(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m2_m))) +vfloat32m2_t vfslide1down(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m4))) +vfloat32m4_t vfslide1down(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m4_m))) +vfloat32m4_t vfslide1down(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m8))) +vfloat32m8_t vfslide1down(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m8_m))) +vfloat32m8_t vfslide1down(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32mf2))) +vfloat32mf2_t vfslide1down(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32mf2_m))) +vfloat32mf2_t vfslide1down(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m1))) +vfloat32m1_t vrgather(vfloat32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m1_m))) +vfloat32m1_t vrgather(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m2))) +vfloat32m2_t vrgather(vfloat32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m2_m))) +vfloat32m2_t vrgather(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m4))) +vfloat32m4_t vrgather(vfloat32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m4_m))) +vfloat32m4_t vrgather(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m8))) +vfloat32m8_t vrgather(vfloat32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m8_m))) +vfloat32m8_t vrgather(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32mf2))) +vfloat32mf2_t vrgather(vfloat32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32mf2_m))) +vfloat32mf2_t vrgather(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m1))) +vfloat32m1_t vrgather(vfloat32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m1_m))) +vfloat32m1_t vrgather(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m2))) +vfloat32m2_t vrgather(vfloat32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m2_m))) +vfloat32m2_t vrgather(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m4))) +vfloat32m4_t vrgather(vfloat32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m4_m))) +vfloat32m4_t vrgather(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m8))) +vfloat32m8_t vrgather(vfloat32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m8_m))) +vfloat32m8_t vrgather(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32mf2))) +vfloat32mf2_t vrgather(vfloat32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32mf2_m))) +vfloat32mf2_t vrgather(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m1))) +vfloat32m1_t vrgatherei16(vfloat32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m1_m))) +vfloat32m1_t vrgatherei16(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m2))) +vfloat32m2_t vrgatherei16(vfloat32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m2_m))) +vfloat32m2_t vrgatherei16(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m4))) +vfloat32m4_t vrgatherei16(vfloat32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m4_m))) +vfloat32m4_t vrgatherei16(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m8))) +vfloat32m8_t vrgatherei16(vfloat32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m8_m))) +vfloat32m8_t vrgatherei16(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32mf2))) +vfloat32mf2_t vrgatherei16(vfloat32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32mf2_m))) +vfloat32mf2_t vrgatherei16(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f32m1))) +vfloat32m1_t vcompress(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f32m2))) +vfloat32m2_t vcompress(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f32m4))) +vfloat32m4_t vcompress(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f32m8))) +vfloat32m8_t vcompress(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f32mf2))) +vfloat32mf2_t vcompress(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f32m2_f32m1))) +vfloat32m1_t vget_f32m1(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f32m4_f32m1))) +vfloat32m1_t vget_f32m1(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f32m8_f32m1))) +vfloat32m1_t vget_f32m1(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f32m4_f32m2))) +vfloat32m2_t vget_f32m2(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f32m8_f32m2))) +vfloat32m2_t vget_f32m2(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f32m8_f32m4))) +vfloat32m4_t vget_f32m4(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f32m1_f32m2))) +vfloat32m2_t vset(vfloat32m2_t op0, size_t op1, vfloat32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f32m1_f32m4))) +vfloat32m4_t vset(vfloat32m4_t op0, size_t op1, vfloat32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f32m2_f32m4))) +vfloat32m4_t vset(vfloat32m4_t op0, size_t op1, vfloat32m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m1))) +void vsuxei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m1_m))) +void vsuxei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m2))) +void vsuxei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m2_m))) +void vsuxei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m4))) +void vsuxei8(float * op0, vuint8m1_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m4_m))) +void vsuxei8(vbool8_t op0, float * op1, vuint8m1_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m8))) +void vsuxei8(float * op0, vuint8m2_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m8_m))) +void vsuxei8(vbool4_t op0, float * op1, vuint8m2_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32mf2))) +void vsuxei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32mf2_m))) +void vsuxei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f32m1_f32m8))) +vfloat32m8_t vset(vfloat32m8_t op0, size_t op1, vfloat32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f32m2_f32m8))) +vfloat32m8_t vset(vfloat32m8_t op0, size_t op1, vfloat32m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f32m4_f32m8))) +vfloat32m8_t vset(vfloat32m8_t op0, size_t op1, vfloat32m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m1))) +void vsuxei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m1_m))) +void vsuxei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m2))) +void vsuxei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m2_m))) +void vsuxei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m4))) +void vsuxei16(float * op0, vuint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m4_m))) +void vsuxei16(vbool8_t op0, float * op1, vuint16m2_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m8))) +void vsuxei16(float * op0, vuint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m8_m))) +void vsuxei16(vbool4_t op0, float * op1, vuint16m4_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32mf2))) +void vsuxei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32mf2_m))) +void vsuxei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m1))) +void vsuxei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m1_m))) +void vsuxei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m2))) +void vsuxei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m2_m))) +void vsuxei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m4))) +void vsuxei32(float * op0, vuint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m4_m))) +void vsuxei32(vbool8_t op0, float * op1, vuint32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m8))) +void vsuxei32(float * op0, vuint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m8_m))) +void vsuxei32(vbool4_t op0, float * op1, vuint32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32mf2))) +void vsuxei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32mf2_m))) +void vsuxei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32m1))) +void vsuxei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32m1_m))) +void vsuxei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32m2))) +void vsuxei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32m2_m))) +void vsuxei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32m4))) +void vsuxei64(float * op0, vuint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32m4_m))) +void vsuxei64(vbool8_t op0, float * op1, vuint64m8_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32mf2))) +void vsuxei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32mf2_m))) +void vsuxei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m1))) +void vsoxei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m1_m))) +void vsoxei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m2))) +void vsoxei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m2_m))) +void vsoxei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m4))) +void vsoxei8(float * op0, vuint8m1_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m4_m))) +void vsoxei8(vbool8_t op0, float * op1, vuint8m1_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m8))) +void vsoxei8(float * op0, vuint8m2_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m8_m))) +void vsoxei8(vbool4_t op0, float * op1, vuint8m2_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32mf2))) +void vsoxei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32mf2_m))) +void vsoxei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m1))) +void vsoxei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m1_m))) +void vsoxei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m2))) +void vsoxei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m2_m))) +void vsoxei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m4))) +void vsoxei16(float * op0, vuint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m4_m))) +void vsoxei16(vbool8_t op0, float * op1, vuint16m2_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m8))) +void vsoxei16(float * op0, vuint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m8_m))) +void vsoxei16(vbool4_t op0, float * op1, vuint16m4_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32mf2))) +void vsoxei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32mf2_m))) +void vsoxei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m1))) +void vsoxei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m1_m))) +void vsoxei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m2))) +void vsoxei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m2_m))) +void vsoxei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m4))) +void vsoxei32(float * op0, vuint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m4_m))) +void vsoxei32(vbool8_t op0, float * op1, vuint32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m8))) +void vsoxei32(float * op0, vuint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m8_m))) +void vsoxei32(vbool4_t op0, float * op1, vuint32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32mf2))) +void vsoxei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32mf2_m))) +void vsoxei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32m1))) +void vsoxei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32m1_m))) +void vsoxei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32m2))) +void vsoxei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32m2_m))) +void vsoxei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32m4))) +void vsoxei64(float * op0, vuint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32m4_m))) +void vsoxei64(vbool8_t op0, float * op1, vuint64m8_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32mf2))) +void vsoxei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32mf2_m))) +void vsoxei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_f32m1_m))) +vfloat32m1_t vle32ff(vbool32_t op0, vfloat32m1_t op1, const float * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_f32m2_m))) +vfloat32m2_t vle32ff(vbool16_t op0, vfloat32m2_t op1, const float * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_f32m4_m))) +vfloat32m4_t vle32ff(vbool8_t op0, vfloat32m4_t op1, const float * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_f32m8_m))) +vfloat32m8_t vle32ff(vbool4_t op0, vfloat32m8_t op1, const float * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_f32mf2_m))) +vfloat32mf2_t vle32ff(vbool64_t op0, vfloat32mf2_t op1, const float * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m1))) +vfloat32m1_t vfneg(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m1_m))) +vfloat32m1_t vfneg(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m2))) +vfloat32m2_t vfneg(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m2_m))) +vfloat32m2_t vfneg(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m4))) +vfloat32m4_t vfneg(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m4_m))) +vfloat32m4_t vfneg(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m8))) +vfloat32m8_t vfneg(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m8_m))) +vfloat32m8_t vfneg(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32mf2))) +vfloat32mf2_t vfneg(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32mf2_m))) +vfloat32mf2_t vfneg(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_f32m1_m))) +vfloat32m1_t vle32(vbool32_t op0, vfloat32m1_t op1, const float * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_f32m2_m))) +vfloat32m2_t vle32(vbool16_t op0, vfloat32m2_t op1, const float * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_f32m4_m))) +vfloat32m4_t vle32(vbool8_t op0, vfloat32m4_t op1, const float * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_f32m8_m))) +vfloat32m8_t vle32(vbool4_t op0, vfloat32m8_t op1, const float * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_f32mf2_m))) +vfloat32mf2_t vle32(vbool64_t op0, vfloat32mf2_t op1, const float * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m1))) +void vse32(float * op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m1_m))) +void vse32(vbool32_t op0, float * op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m2))) +void vse32(float * op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m2_m))) +void vse32(vbool16_t op0, float * op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m4))) +void vse32(float * op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m4_m))) +void vse32(vbool8_t op0, float * op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m8))) +void vse32(float * op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m8_m))) +void vse32(vbool4_t op0, float * op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32mf2))) +void vse32(float * op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32mf2_m))) +void vse32(vbool64_t op0, float * op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_f32m1_m))) +vfloat32m1_t vlse32(vbool32_t op0, vfloat32m1_t op1, const float * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_f32m2_m))) +vfloat32m2_t vlse32(vbool16_t op0, vfloat32m2_t op1, const float * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_f32m4_m))) +vfloat32m4_t vlse32(vbool8_t op0, vfloat32m4_t op1, const float * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_f32m8_m))) +vfloat32m8_t vlse32(vbool4_t op0, vfloat32m8_t op1, const float * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_f32mf2_m))) +vfloat32mf2_t vlse32(vbool64_t op0, vfloat32mf2_t op1, const float * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m1))) +vuint32m1_t vfclass(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m1_m))) +vuint32m1_t vfclass(vbool32_t op0, vuint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m2))) +vuint32m2_t vfclass(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m2_m))) +vuint32m2_t vfclass(vbool16_t op0, vuint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m4))) +vuint32m4_t vfclass(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m4_m))) +vuint32m4_t vfclass(vbool8_t op0, vuint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m8))) +vuint32m8_t vfclass(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m8_m))) +vuint32m8_t vfclass(vbool4_t op0, vuint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32mf2))) +vuint32mf2_t vfclass(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32mf2_m))) +vuint32mf2_t vfclass(vbool64_t op0, vuint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m1))) +vfloat32m1_t vfcvt_f(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m1_m))) +vfloat32m1_t vfcvt_f(vbool32_t op0, vfloat32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m2))) +vfloat32m2_t vfcvt_f(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m2_m))) +vfloat32m2_t vfcvt_f(vbool16_t op0, vfloat32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m4))) +vfloat32m4_t vfcvt_f(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m4_m))) +vfloat32m4_t vfcvt_f(vbool8_t op0, vfloat32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m8))) +vfloat32m8_t vfcvt_f(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m8_m))) +vfloat32m8_t vfcvt_f(vbool4_t op0, vfloat32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32mf2))) +vfloat32mf2_t vfcvt_f(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32mf2_m))) +vfloat32mf2_t vfcvt_f(vbool64_t op0, vfloat32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m1))) +vfloat32m1_t vfcvt_f(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m1_m))) +vfloat32m1_t vfcvt_f(vbool32_t op0, vfloat32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m2))) +vfloat32m2_t vfcvt_f(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m2_m))) +vfloat32m2_t vfcvt_f(vbool16_t op0, vfloat32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m4))) +vfloat32m4_t vfcvt_f(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m4_m))) +vfloat32m4_t vfcvt_f(vbool8_t op0, vfloat32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m8))) +vfloat32m8_t vfcvt_f(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m8_m))) +vfloat32m8_t vfcvt_f(vbool4_t op0, vfloat32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32mf2))) +vfloat32mf2_t vfcvt_f(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32mf2_m))) +vfloat32mf2_t vfcvt_f(vbool64_t op0, vfloat32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m1))) +vint32m1_t vfcvt_rtz_x(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m1_m))) +vint32m1_t vfcvt_rtz_x(vbool32_t op0, vint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m2))) +vint32m2_t vfcvt_rtz_x(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m2_m))) +vint32m2_t vfcvt_rtz_x(vbool16_t op0, vint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m4))) +vint32m4_t vfcvt_rtz_x(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m4_m))) +vint32m4_t vfcvt_rtz_x(vbool8_t op0, vint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m8))) +vint32m8_t vfcvt_rtz_x(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m8_m))) +vint32m8_t vfcvt_rtz_x(vbool4_t op0, vint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32mf2))) +vint32mf2_t vfcvt_rtz_x(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32mf2_m))) +vint32mf2_t vfcvt_rtz_x(vbool64_t op0, vint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m1))) +vuint32m1_t vfcvt_rtz_xu(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m1_m))) +vuint32m1_t vfcvt_rtz_xu(vbool32_t op0, vuint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m2))) +vuint32m2_t vfcvt_rtz_xu(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m2_m))) +vuint32m2_t vfcvt_rtz_xu(vbool16_t op0, vuint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m4))) +vuint32m4_t vfcvt_rtz_xu(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m4_m))) +vuint32m4_t vfcvt_rtz_xu(vbool8_t op0, vuint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m8))) +vuint32m8_t vfcvt_rtz_xu(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m8_m))) +vuint32m8_t vfcvt_rtz_xu(vbool4_t op0, vuint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32mf2))) +vuint32mf2_t vfcvt_rtz_xu(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32mf2_m))) +vuint32mf2_t vfcvt_rtz_xu(vbool64_t op0, vuint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m1))) +vint32m1_t vfcvt_x(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m1_m))) +vint32m1_t vfcvt_x(vbool32_t op0, vint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m2))) +vint32m2_t vfcvt_x(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m2_m))) +vint32m2_t vfcvt_x(vbool16_t op0, vint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m4))) +vint32m4_t vfcvt_x(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m4_m))) +vint32m4_t vfcvt_x(vbool8_t op0, vint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m8))) +vint32m8_t vfcvt_x(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m8_m))) +vint32m8_t vfcvt_x(vbool4_t op0, vint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32mf2))) +vint32mf2_t vfcvt_x(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32mf2_m))) +vint32mf2_t vfcvt_x(vbool64_t op0, vint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m1))) +vuint32m1_t vfcvt_xu(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m1_m))) +vuint32m1_t vfcvt_xu(vbool32_t op0, vuint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m2))) +vuint32m2_t vfcvt_xu(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m2_m))) +vuint32m2_t vfcvt_xu(vbool16_t op0, vuint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m4))) +vuint32m4_t vfcvt_xu(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m4_m))) +vuint32m4_t vfcvt_xu(vbool8_t op0, vuint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m8))) +vuint32m8_t vfcvt_xu(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m8_m))) +vuint32m8_t vfcvt_xu(vbool4_t op0, vuint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32mf2))) +vuint32mf2_t vfcvt_xu(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32mf2_m))) +vuint32mf2_t vfcvt_xu(vbool64_t op0, vuint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32mf2))) +vfloat32mf2_t vfncvt_f(vint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32mf2_m))) +vfloat32mf2_t vfncvt_f(vbool64_t op0, vfloat32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32m1))) +vfloat32m1_t vfncvt_f(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32m1_m))) +vfloat32m1_t vfncvt_f(vbool32_t op0, vfloat32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32m2))) +vfloat32m2_t vfncvt_f(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32m2_m))) +vfloat32m2_t vfncvt_f(vbool16_t op0, vfloat32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32m4))) +vfloat32m4_t vfncvt_f(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32m4_m))) +vfloat32m4_t vfncvt_f(vbool8_t op0, vfloat32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32mf2))) +vfloat32mf2_t vfncvt_f(vuint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32mf2_m))) +vfloat32mf2_t vfncvt_f(vbool64_t op0, vfloat32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32m1))) +vfloat32m1_t vfncvt_f(vuint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32m1_m))) +vfloat32m1_t vfncvt_f(vbool32_t op0, vfloat32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32m2))) +vfloat32m2_t vfncvt_f(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32m2_m))) +vfloat32m2_t vfncvt_f(vbool16_t op0, vfloat32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32m4))) +vfloat32m4_t vfncvt_f(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32m4_m))) +vfloat32m4_t vfncvt_f(vbool8_t op0, vfloat32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16mf4))) +vint16mf4_t vfncvt_rtz_x(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16mf4_m))) +vint16mf4_t vfncvt_rtz_x(vbool64_t op0, vint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16mf2))) +vint16mf2_t vfncvt_rtz_x(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16mf2_m))) +vint16mf2_t vfncvt_rtz_x(vbool32_t op0, vint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16m1))) +vint16m1_t vfncvt_rtz_x(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16m1_m))) +vint16m1_t vfncvt_rtz_x(vbool16_t op0, vint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16m2))) +vint16m2_t vfncvt_rtz_x(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16m2_m))) +vint16m2_t vfncvt_rtz_x(vbool8_t op0, vint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16m4))) +vint16m4_t vfncvt_rtz_x(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16m4_m))) +vint16m4_t vfncvt_rtz_x(vbool4_t op0, vint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf4))) +vuint16mf4_t vfncvt_rtz_xu(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf4_m))) +vuint16mf4_t vfncvt_rtz_xu(vbool64_t op0, vuint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf2))) +vuint16mf2_t vfncvt_rtz_xu(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf2_m))) +vuint16mf2_t vfncvt_rtz_xu(vbool32_t op0, vuint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16m1))) +vuint16m1_t vfncvt_rtz_xu(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16m1_m))) +vuint16m1_t vfncvt_rtz_xu(vbool16_t op0, vuint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16m2))) +vuint16m2_t vfncvt_rtz_xu(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16m2_m))) +vuint16m2_t vfncvt_rtz_xu(vbool8_t op0, vuint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16m4))) +vuint16m4_t vfncvt_rtz_xu(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16m4_m))) +vuint16m4_t vfncvt_rtz_xu(vbool4_t op0, vuint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16mf4))) +vint16mf4_t vfncvt_x(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16mf4_m))) +vint16mf4_t vfncvt_x(vbool64_t op0, vint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16mf2))) +vint16mf2_t vfncvt_x(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16mf2_m))) +vint16mf2_t vfncvt_x(vbool32_t op0, vint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16m1))) +vint16m1_t vfncvt_x(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16m1_m))) +vint16m1_t vfncvt_x(vbool16_t op0, vint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16m2))) +vint16m2_t vfncvt_x(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16m2_m))) +vint16m2_t vfncvt_x(vbool8_t op0, vint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16m4))) +vint16m4_t vfncvt_x(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16m4_m))) +vint16m4_t vfncvt_x(vbool4_t op0, vint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16mf4))) +vuint16mf4_t vfncvt_xu(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16mf4_m))) +vuint16mf4_t vfncvt_xu(vbool64_t op0, vuint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16mf2))) +vuint16mf2_t vfncvt_xu(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16mf2_m))) +vuint16mf2_t vfncvt_xu(vbool32_t op0, vuint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16m1))) +vuint16m1_t vfncvt_xu(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16m1_m))) +vuint16m1_t vfncvt_xu(vbool16_t op0, vuint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16m2))) +vuint16m2_t vfncvt_xu(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16m2_m))) +vuint16m2_t vfncvt_xu(vbool8_t op0, vuint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16m4))) +vuint16m4_t vfncvt_xu(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16m4_m))) +vuint16m4_t vfncvt_xu(vbool4_t op0, vuint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m1))) +vfloat32m1_t vfrec7(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m1_m))) +vfloat32m1_t vfrec7(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m2))) +vfloat32m2_t vfrec7(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m2_m))) +vfloat32m2_t vfrec7(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m4))) +vfloat32m4_t vfrec7(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m4_m))) +vfloat32m4_t vfrec7(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m8))) +vfloat32m8_t vfrec7(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m8_m))) +vfloat32m8_t vfrec7(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32mf2))) +vfloat32mf2_t vfrec7(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32mf2_m))) +vfloat32mf2_t vfrec7(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m1))) +vfloat32m1_t vfrsqrt7(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m1_m))) +vfloat32m1_t vfrsqrt7(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m2))) +vfloat32m2_t vfrsqrt7(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m2_m))) +vfloat32m2_t vfrsqrt7(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m4))) +vfloat32m4_t vfrsqrt7(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m4_m))) +vfloat32m4_t vfrsqrt7(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m8))) +vfloat32m8_t vfrsqrt7(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m8_m))) +vfloat32m8_t vfrsqrt7(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32mf2))) +vfloat32mf2_t vfrsqrt7(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32mf2_m))) +vfloat32mf2_t vfrsqrt7(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m1))) +vfloat32m1_t vfsqrt(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m1_m))) +vfloat32m1_t vfsqrt(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m2))) +vfloat32m2_t vfsqrt(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m2_m))) +vfloat32m2_t vfsqrt(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m4))) +vfloat32m4_t vfsqrt(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m4_m))) +vfloat32m4_t vfsqrt(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m8))) +vfloat32m8_t vfsqrt(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m8_m))) +vfloat32m8_t vfsqrt(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32mf2))) +vfloat32mf2_t vfsqrt(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32mf2_m))) +vfloat32mf2_t vfsqrt(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32mf2))) +vfloat32mf2_t vfwcvt_f(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32mf2_m))) +vfloat32mf2_t vfwcvt_f(vbool64_t op0, vfloat32mf2_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m1))) +vfloat32m1_t vfwcvt_f(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m1_m))) +vfloat32m1_t vfwcvt_f(vbool32_t op0, vfloat32m1_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m2))) +vfloat32m2_t vfwcvt_f(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m2_m))) +vfloat32m2_t vfwcvt_f(vbool16_t op0, vfloat32m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m4))) +vfloat32m4_t vfwcvt_f(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m4_m))) +vfloat32m4_t vfwcvt_f(vbool8_t op0, vfloat32m4_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m8))) +vfloat32m8_t vfwcvt_f(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m8_m))) +vfloat32m8_t vfwcvt_f(vbool4_t op0, vfloat32m8_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32mf2))) +vfloat32mf2_t vfwcvt_f(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32mf2_m))) +vfloat32mf2_t vfwcvt_f(vbool64_t op0, vfloat32mf2_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m1))) +vfloat32m1_t vfwcvt_f(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m1_m))) +vfloat32m1_t vfwcvt_f(vbool32_t op0, vfloat32m1_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m2))) +vfloat32m2_t vfwcvt_f(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m2_m))) +vfloat32m2_t vfwcvt_f(vbool16_t op0, vfloat32m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m4))) +vfloat32m4_t vfwcvt_f(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m4_m))) +vfloat32m4_t vfwcvt_f(vbool8_t op0, vfloat32m4_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m8))) +vfloat32m8_t vfwcvt_f(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m8_m))) +vfloat32m8_t vfwcvt_f(vbool4_t op0, vfloat32m8_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m1))) +vint64m1_t vfwcvt_rtz_x(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m1_m))) +vint64m1_t vfwcvt_rtz_x(vbool64_t op0, vint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m2))) +vint64m2_t vfwcvt_rtz_x(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m2_m))) +vint64m2_t vfwcvt_rtz_x(vbool32_t op0, vint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m4))) +vint64m4_t vfwcvt_rtz_x(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m4_m))) +vint64m4_t vfwcvt_rtz_x(vbool16_t op0, vint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m8))) +vint64m8_t vfwcvt_rtz_x(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m8_m))) +vint64m8_t vfwcvt_rtz_x(vbool8_t op0, vint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m1))) +vuint64m1_t vfwcvt_rtz_xu(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m1_m))) +vuint64m1_t vfwcvt_rtz_xu(vbool64_t op0, vuint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m2))) +vuint64m2_t vfwcvt_rtz_xu(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m2_m))) +vuint64m2_t vfwcvt_rtz_xu(vbool32_t op0, vuint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m4))) +vuint64m4_t vfwcvt_rtz_xu(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m4_m))) +vuint64m4_t vfwcvt_rtz_xu(vbool16_t op0, vuint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m8))) +vuint64m8_t vfwcvt_rtz_xu(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m8_m))) +vuint64m8_t vfwcvt_rtz_xu(vbool8_t op0, vuint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m1))) +vint64m1_t vfwcvt_x(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m1_m))) +vint64m1_t vfwcvt_x(vbool64_t op0, vint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m2))) +vint64m2_t vfwcvt_x(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m2_m))) +vint64m2_t vfwcvt_x(vbool32_t op0, vint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m4))) +vint64m4_t vfwcvt_x(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m4_m))) +vint64m4_t vfwcvt_x(vbool16_t op0, vint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m8))) +vint64m8_t vfwcvt_x(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m8_m))) +vint64m8_t vfwcvt_x(vbool8_t op0, vint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m1))) +vuint64m1_t vfwcvt_xu(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m1_m))) +vuint64m1_t vfwcvt_xu(vbool64_t op0, vuint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m2))) +vuint64m2_t vfwcvt_xu(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m2_m))) +vuint64m2_t vfwcvt_xu(vbool32_t op0, vuint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m4))) +vuint64m4_t vfwcvt_xu(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m4_m))) +vuint64m4_t vfwcvt_xu(vbool16_t op0, vuint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m8))) +vuint64m8_t vfwcvt_xu(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m8_m))) +vuint64m8_t vfwcvt_xu(vbool8_t op0, vuint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32mf2_f32m1))) +vfloat32m1_t vlmul_ext_f32m1(vfloat32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32m1_f32m2))) +vfloat32m2_t vlmul_ext_f32m2(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32mf2_f32m2))) +vfloat32m2_t vlmul_ext_f32m2(vfloat32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32m1_f32m4))) +vfloat32m4_t vlmul_ext_f32m4(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32m2_f32m4))) +vfloat32m4_t vlmul_ext_f32m4(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32mf2_f32m4))) +vfloat32m4_t vlmul_ext_f32m4(vfloat32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32m1_f32m8))) +vfloat32m8_t vlmul_ext_f32m8(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32m2_f32m8))) +vfloat32m8_t vlmul_ext_f32m8(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32m4_f32m8))) +vfloat32m8_t vlmul_ext_f32m8(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32mf2_f32m8))) +vfloat32m8_t vlmul_ext_f32m8(vfloat32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m1_f32mf2))) +vfloat32mf2_t vlmul_trunc_f32mf2(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m2_f32mf2))) +vfloat32mf2_t vlmul_trunc_f32mf2(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m4_f32mf2))) +vfloat32mf2_t vlmul_trunc_f32mf2(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m8_f32mf2))) +vfloat32mf2_t vlmul_trunc_f32mf2(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m2_f32m1))) +vfloat32m1_t vlmul_trunc_f32m1(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m4_f32m1))) +vfloat32m1_t vlmul_trunc_f32m1(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m8_f32m1))) +vfloat32m1_t vlmul_trunc_f32m1(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m4_f32m2))) +vfloat32m2_t vlmul_trunc_f32m2(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m8_f32m2))) +vfloat32m2_t vlmul_trunc_f32m2(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m8_f32m4))) +vfloat32m4_t vlmul_trunc_f32m4(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m1_f32m1))) +vfloat32m1_t vreinterpret_f32m1(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m2_f32m2))) +vfloat32m2_t vreinterpret_f32m2(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m4_f32m4))) +vfloat32m4_t vreinterpret_f32m4(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m8_f32m8))) +vfloat32m8_t vreinterpret_f32m8(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32mf2_f32mf2))) +vfloat32mf2_t vreinterpret_f32mf2(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m1_f32m1))) +vfloat32m1_t vreinterpret_f32m1(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m2_f32m2))) +vfloat32m2_t vreinterpret_f32m2(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m4_f32m4))) +vfloat32m4_t vreinterpret_f32m4(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m8_f32m8))) +vfloat32m8_t vreinterpret_f32m8(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32mf2_f32mf2))) +vfloat32mf2_t vreinterpret_f32mf2(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m1_i32m1))) +vint32m1_t vreinterpret_i32m1(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m2_i32m2))) +vint32m2_t vreinterpret_i32m2(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m4_i32m4))) +vint32m4_t vreinterpret_i32m4(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m8_i32m8))) +vint32m8_t vreinterpret_i32m8(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32mf2_i32mf2))) +vint32mf2_t vreinterpret_i32mf2(vfloat32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m1_u32m1))) +vuint32m1_t vreinterpret_u32m1(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m2_u32m2))) +vuint32m2_t vreinterpret_u32m2(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m4_u32m4))) +vuint32m4_t vreinterpret_u32m4(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m8_u32m8))) +vuint32m8_t vreinterpret_u32m8(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32mf2_u32mf2))) +vuint32mf2_t vreinterpret_u32mf2(vfloat32mf2_t op0); + +#endif + +#if defined(__riscv_d) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_f64m1_m))) +vfloat64m1_t vlse64(vbool64_t op0, vfloat64m1_t op1, const double * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_f64m2_m))) +vfloat64m2_t vlse64(vbool32_t op0, vfloat64m2_t op1, const double * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_f64m4_m))) +vfloat64m4_t vlse64(vbool16_t op0, vfloat64m4_t op1, const double * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_f64m8_m))) +vfloat64m8_t vlse64(vbool8_t op0, vfloat64m8_t op1, const double * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m1))) +void vsse64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m1_m))) +void vsse64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m2))) +void vsse64(double * op0, ptrdiff_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m2_m))) +void vsse64(vbool32_t op0, double * op1, ptrdiff_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m4))) +void vsse64(double * op0, ptrdiff_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m4_m))) +void vsse64(vbool16_t op0, double * op1, ptrdiff_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m8))) +void vsse64(double * op0, ptrdiff_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m8_m))) +void vsse64(vbool8_t op0, double * op1, ptrdiff_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m1))) +vfloat64m1_t vluxei8(const double * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m1_m))) +vfloat64m1_t vluxei8(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m2))) +vfloat64m2_t vluxei8(const double * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m2_m))) +vfloat64m2_t vluxei8(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m4))) +vfloat64m4_t vluxei8(const double * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m4_m))) +vfloat64m4_t vluxei8(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m8))) +vfloat64m8_t vluxei8(const double * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m8_m))) +vfloat64m8_t vluxei8(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m1))) +vfloat64m1_t vluxei16(const double * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m1_m))) +vfloat64m1_t vluxei16(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m2))) +vfloat64m2_t vluxei16(const double * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m2_m))) +vfloat64m2_t vluxei16(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m4))) +vfloat64m4_t vluxei16(const double * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m4_m))) +vfloat64m4_t vluxei16(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m8))) +vfloat64m8_t vluxei16(const double * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m8_m))) +vfloat64m8_t vluxei16(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m1))) +vfloat64m1_t vluxei32(const double * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m1_m))) +vfloat64m1_t vluxei32(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m2))) +vfloat64m2_t vluxei32(const double * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m2_m))) +vfloat64m2_t vluxei32(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m4))) +vfloat64m4_t vluxei32(const double * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m4_m))) +vfloat64m4_t vluxei32(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m8))) +vfloat64m8_t vluxei32(const double * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m8_m))) +vfloat64m8_t vluxei32(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m1))) +vfloat64m1_t vluxei64(const double * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m1_m))) +vfloat64m1_t vluxei64(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m2))) +vfloat64m2_t vluxei64(const double * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m2_m))) +vfloat64m2_t vluxei64(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m4))) +vfloat64m4_t vluxei64(const double * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m4_m))) +vfloat64m4_t vluxei64(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m8))) +vfloat64m8_t vluxei64(const double * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m8_m))) +vfloat64m8_t vluxei64(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m1))) +vfloat64m1_t vloxei8(const double * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m1_m))) +vfloat64m1_t vloxei8(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m2))) +vfloat64m2_t vloxei8(const double * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m2_m))) +vfloat64m2_t vloxei8(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m4))) +vfloat64m4_t vloxei8(const double * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m4_m))) +vfloat64m4_t vloxei8(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m8))) +vfloat64m8_t vloxei8(const double * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m8_m))) +vfloat64m8_t vloxei8(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m1))) +vfloat64m1_t vloxei16(const double * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m1_m))) +vfloat64m1_t vloxei16(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m2))) +vfloat64m2_t vloxei16(const double * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m2_m))) +vfloat64m2_t vloxei16(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m4))) +vfloat64m4_t vloxei16(const double * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m4_m))) +vfloat64m4_t vloxei16(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m8))) +vfloat64m8_t vloxei16(const double * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m8_m))) +vfloat64m8_t vloxei16(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m1))) +vfloat64m1_t vloxei32(const double * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m1_m))) +vfloat64m1_t vloxei32(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m2))) +vfloat64m2_t vloxei32(const double * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m2_m))) +vfloat64m2_t vloxei32(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m4))) +vfloat64m4_t vloxei32(const double * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m4_m))) +vfloat64m4_t vloxei32(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m8))) +vfloat64m8_t vloxei32(const double * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m8_m))) +vfloat64m8_t vloxei32(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m1))) +vfloat64m1_t vloxei64(const double * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m1_m))) +vfloat64m1_t vloxei64(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m2))) +vfloat64m2_t vloxei64(const double * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m2_m))) +vfloat64m2_t vloxei64(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m4))) +vfloat64m4_t vloxei64(const double * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m4_m))) +vfloat64m4_t vloxei64(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m8))) +vfloat64m8_t vloxei64(const double * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m8_m))) +vfloat64m8_t vloxei64(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f64m1))) +vfloat64m1_t vmv_v(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f64m2))) +vfloat64m2_t vmv_v(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f64m4))) +vfloat64m4_t vmv_v(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f64m8))) +vfloat64m8_t vmv_v(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m1))) +vfloat64m1_t vfadd(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m1_m))) +vfloat64m1_t vfadd(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m2))) +vfloat64m2_t vfadd(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m2_m))) +vfloat64m2_t vfadd(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m4))) +vfloat64m4_t vfadd(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m4_m))) +vfloat64m4_t vfadd(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m8))) +vfloat64m8_t vfadd(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m8_m))) +vfloat64m8_t vfadd(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m1))) +vfloat64m1_t vfadd(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m1_m))) +vfloat64m1_t vfadd(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m2))) +vfloat64m2_t vfadd(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m2_m))) +vfloat64m2_t vfadd(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m4))) +vfloat64m4_t vfadd(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m4_m))) +vfloat64m4_t vfadd(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m8))) +vfloat64m8_t vfadd(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m8_m))) +vfloat64m8_t vfadd(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m1))) +vfloat64m1_t vfsub(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m1_m))) +vfloat64m1_t vfsub(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m2))) +vfloat64m2_t vfsub(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m2_m))) +vfloat64m2_t vfsub(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m4))) +vfloat64m4_t vfsub(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m4_m))) +vfloat64m4_t vfsub(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m8))) +vfloat64m8_t vfsub(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m8_m))) +vfloat64m8_t vfsub(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m1))) +vfloat64m1_t vfsub(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m1_m))) +vfloat64m1_t vfsub(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m2))) +vfloat64m2_t vfsub(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m2_m))) +vfloat64m2_t vfsub(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m4))) +vfloat64m4_t vfsub(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m4_m))) +vfloat64m4_t vfsub(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m8))) +vfloat64m8_t vfsub(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m8_m))) +vfloat64m8_t vfsub(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m1))) +vfloat64m1_t vfrsub(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m1_m))) +vfloat64m1_t vfrsub(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m2))) +vfloat64m2_t vfrsub(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m2_m))) +vfloat64m2_t vfrsub(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m4))) +vfloat64m4_t vfrsub(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m4_m))) +vfloat64m4_t vfrsub(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m8))) +vfloat64m8_t vfrsub(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m8_m))) +vfloat64m8_t vfrsub(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m1))) +vfloat64m1_t vfmul(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m1_m))) +vfloat64m1_t vfmul(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m2))) +vfloat64m2_t vfmul(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m2_m))) +vfloat64m2_t vfmul(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m4))) +vfloat64m4_t vfmul(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m4_m))) +vfloat64m4_t vfmul(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m8))) +vfloat64m8_t vfmul(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m8_m))) +vfloat64m8_t vfmul(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m1))) +vfloat64m1_t vfmul(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m1_m))) +vfloat64m1_t vfmul(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m2))) +vfloat64m2_t vfmul(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m2_m))) +vfloat64m2_t vfmul(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m4))) +vfloat64m4_t vfmul(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m4_m))) +vfloat64m4_t vfmul(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m8))) +vfloat64m8_t vfmul(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m8_m))) +vfloat64m8_t vfmul(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m1))) +vfloat64m1_t vfdiv(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m1_m))) +vfloat64m1_t vfdiv(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m2))) +vfloat64m2_t vfdiv(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m2_m))) +vfloat64m2_t vfdiv(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m4))) +vfloat64m4_t vfdiv(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m4_m))) +vfloat64m4_t vfdiv(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m8))) +vfloat64m8_t vfdiv(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m8_m))) +vfloat64m8_t vfdiv(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m1))) +vfloat64m1_t vfdiv(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m1_m))) +vfloat64m1_t vfdiv(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m2))) +vfloat64m2_t vfdiv(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m2_m))) +vfloat64m2_t vfdiv(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m4))) +vfloat64m4_t vfdiv(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m4_m))) +vfloat64m4_t vfdiv(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m8))) +vfloat64m8_t vfdiv(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m8_m))) +vfloat64m8_t vfdiv(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m1))) +vfloat64m1_t vfrdiv(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m1_m))) +vfloat64m1_t vfrdiv(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m2))) +vfloat64m2_t vfrdiv(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m2_m))) +vfloat64m2_t vfrdiv(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m4))) +vfloat64m4_t vfrdiv(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m4_m))) +vfloat64m4_t vfrdiv(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m8))) +vfloat64m8_t vfrdiv(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m8_m))) +vfloat64m8_t vfrdiv(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m1))) +vfloat64m1_t vfmacc(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m1_m))) +vfloat64m1_t vfmacc(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m2))) +vfloat64m2_t vfmacc(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m2_m))) +vfloat64m2_t vfmacc(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m4))) +vfloat64m4_t vfmacc(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m4_m))) +vfloat64m4_t vfmacc(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m8))) +vfloat64m8_t vfmacc(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m8_m))) +vfloat64m8_t vfmacc(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m1))) +vfloat64m1_t vfmacc(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m1_m))) +vfloat64m1_t vfmacc(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m2))) +vfloat64m2_t vfmacc(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m2_m))) +vfloat64m2_t vfmacc(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m4))) +vfloat64m4_t vfmacc(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m4_m))) +vfloat64m4_t vfmacc(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m8))) +vfloat64m8_t vfmacc(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m8_m))) +vfloat64m8_t vfmacc(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m1))) +vfloat64m1_t vfnmacc(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m1_m))) +vfloat64m1_t vfnmacc(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m2))) +vfloat64m2_t vfnmacc(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m2_m))) +vfloat64m2_t vfnmacc(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m4))) +vfloat64m4_t vfnmacc(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m4_m))) +vfloat64m4_t vfnmacc(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m8))) +vfloat64m8_t vfnmacc(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m8_m))) +vfloat64m8_t vfnmacc(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m1))) +vfloat64m1_t vfnmacc(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m1_m))) +vfloat64m1_t vfnmacc(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m2))) +vfloat64m2_t vfnmacc(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m2_m))) +vfloat64m2_t vfnmacc(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m4))) +vfloat64m4_t vfnmacc(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m4_m))) +vfloat64m4_t vfnmacc(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m8))) +vfloat64m8_t vfnmacc(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m8_m))) +vfloat64m8_t vfnmacc(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m1))) +vfloat64m1_t vfmsac(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m1_m))) +vfloat64m1_t vfmsac(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m2))) +vfloat64m2_t vfmsac(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m2_m))) +vfloat64m2_t vfmsac(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m4))) +vfloat64m4_t vfmsac(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m4_m))) +vfloat64m4_t vfmsac(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m8))) +vfloat64m8_t vfmsac(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m8_m))) +vfloat64m8_t vfmsac(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m1))) +vfloat64m1_t vfmsac(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m1_m))) +vfloat64m1_t vfmsac(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m2))) +vfloat64m2_t vfmsac(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m2_m))) +vfloat64m2_t vfmsac(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m4))) +vfloat64m4_t vfmsac(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m4_m))) +vfloat64m4_t vfmsac(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m8))) +vfloat64m8_t vfmsac(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m8_m))) +vfloat64m8_t vfmsac(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m1))) +vfloat64m1_t vfnmsac(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m1_m))) +vfloat64m1_t vfnmsac(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m2))) +vfloat64m2_t vfnmsac(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m2_m))) +vfloat64m2_t vfnmsac(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m4))) +vfloat64m4_t vfnmsac(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m4_m))) +vfloat64m4_t vfnmsac(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m8))) +vfloat64m8_t vfnmsac(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m8_m))) +vfloat64m8_t vfnmsac(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m1))) +vfloat64m1_t vfnmsac(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m1_m))) +vfloat64m1_t vfnmsac(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m2))) +vfloat64m2_t vfnmsac(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m2_m))) +vfloat64m2_t vfnmsac(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m4))) +vfloat64m4_t vfnmsac(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m4_m))) +vfloat64m4_t vfnmsac(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m8))) +vfloat64m8_t vfnmsac(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m8_m))) +vfloat64m8_t vfnmsac(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m1))) +vfloat64m1_t vfmadd(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m1_m))) +vfloat64m1_t vfmadd(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m2))) +vfloat64m2_t vfmadd(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m2_m))) +vfloat64m2_t vfmadd(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m4))) +vfloat64m4_t vfmadd(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m4_m))) +vfloat64m4_t vfmadd(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m8))) +vfloat64m8_t vfmadd(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m8_m))) +vfloat64m8_t vfmadd(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m1))) +vfloat64m1_t vfmadd(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m1_m))) +vfloat64m1_t vfmadd(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m2))) +vfloat64m2_t vfmadd(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m2_m))) +vfloat64m2_t vfmadd(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m4))) +vfloat64m4_t vfmadd(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m4_m))) +vfloat64m4_t vfmadd(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m8))) +vfloat64m8_t vfmadd(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m8_m))) +vfloat64m8_t vfmadd(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m1))) +vfloat64m1_t vfnmadd(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m1_m))) +vfloat64m1_t vfnmadd(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m2))) +vfloat64m2_t vfnmadd(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m2_m))) +vfloat64m2_t vfnmadd(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m4))) +vfloat64m4_t vfnmadd(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m4_m))) +vfloat64m4_t vfnmadd(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m8))) +vfloat64m8_t vfnmadd(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m8_m))) +vfloat64m8_t vfnmadd(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m1))) +vfloat64m1_t vfnmadd(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m1_m))) +vfloat64m1_t vfnmadd(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m2))) +vfloat64m2_t vfnmadd(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m2_m))) +vfloat64m2_t vfnmadd(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m4))) +vfloat64m4_t vfnmadd(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m4_m))) +vfloat64m4_t vfnmadd(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m8))) +vfloat64m8_t vfnmadd(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m8_m))) +vfloat64m8_t vfnmadd(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m1))) +vfloat64m1_t vfmsub(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m1_m))) +vfloat64m1_t vfmsub(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m2))) +vfloat64m2_t vfmsub(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m2_m))) +vfloat64m2_t vfmsub(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m4))) +vfloat64m4_t vfmsub(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m4_m))) +vfloat64m4_t vfmsub(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m8))) +vfloat64m8_t vfmsub(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m8_m))) +vfloat64m8_t vfmsub(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m1))) +vfloat64m1_t vfmsub(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m1_m))) +vfloat64m1_t vfmsub(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m2))) +vfloat64m2_t vfmsub(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m2_m))) +vfloat64m2_t vfmsub(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m4))) +vfloat64m4_t vfmsub(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m4_m))) +vfloat64m4_t vfmsub(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m8))) +vfloat64m8_t vfmsub(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m8_m))) +vfloat64m8_t vfmsub(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m1))) +vfloat64m1_t vfnmsub(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m1_m))) +vfloat64m1_t vfnmsub(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m2))) +vfloat64m2_t vfnmsub(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m2_m))) +vfloat64m2_t vfnmsub(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m4))) +vfloat64m4_t vfnmsub(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m4_m))) +vfloat64m4_t vfnmsub(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m8))) +vfloat64m8_t vfnmsub(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m8_m))) +vfloat64m8_t vfnmsub(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m1))) +vfloat64m1_t vfnmsub(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m1_m))) +vfloat64m1_t vfnmsub(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m2))) +vfloat64m2_t vfnmsub(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m2_m))) +vfloat64m2_t vfnmsub(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m4))) +vfloat64m4_t vfnmsub(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m4_m))) +vfloat64m4_t vfnmsub(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m8))) +vfloat64m8_t vfnmsub(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m8_m))) +vfloat64m8_t vfnmsub(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m1))) +vfloat64m1_t vfmin(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m1_m))) +vfloat64m1_t vfmin(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m2))) +vfloat64m2_t vfmin(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m2_m))) +vfloat64m2_t vfmin(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m4))) +vfloat64m4_t vfmin(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m4_m))) +vfloat64m4_t vfmin(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m8))) +vfloat64m8_t vfmin(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m8_m))) +vfloat64m8_t vfmin(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m1))) +vfloat64m1_t vfmin(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m1_m))) +vfloat64m1_t vfmin(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m2))) +vfloat64m2_t vfmin(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m2_m))) +vfloat64m2_t vfmin(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m4))) +vfloat64m4_t vfmin(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m4_m))) +vfloat64m4_t vfmin(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m8))) +vfloat64m8_t vfmin(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m8_m))) +vfloat64m8_t vfmin(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m1))) +vfloat64m1_t vfmax(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m1_m))) +vfloat64m1_t vfmax(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m2))) +vfloat64m2_t vfmax(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m2_m))) +vfloat64m2_t vfmax(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m4))) +vfloat64m4_t vfmax(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m4_m))) +vfloat64m4_t vfmax(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m8))) +vfloat64m8_t vfmax(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m8_m))) +vfloat64m8_t vfmax(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m1))) +vfloat64m1_t vfmax(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m1_m))) +vfloat64m1_t vfmax(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m2))) +vfloat64m2_t vfmax(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m2_m))) +vfloat64m2_t vfmax(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m4))) +vfloat64m4_t vfmax(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m4_m))) +vfloat64m4_t vfmax(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m8))) +vfloat64m8_t vfmax(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m8_m))) +vfloat64m8_t vfmax(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m1))) +vfloat64m1_t vfsgnj(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m1_m))) +vfloat64m1_t vfsgnj(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m2))) +vfloat64m2_t vfsgnj(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m2_m))) +vfloat64m2_t vfsgnj(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m4))) +vfloat64m4_t vfsgnj(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m4_m))) +vfloat64m4_t vfsgnj(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m8))) +vfloat64m8_t vfsgnj(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m8_m))) +vfloat64m8_t vfsgnj(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m1))) +vfloat64m1_t vfsgnj(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m1_m))) +vfloat64m1_t vfsgnj(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m2))) +vfloat64m2_t vfsgnj(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m2_m))) +vfloat64m2_t vfsgnj(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m4))) +vfloat64m4_t vfsgnj(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m4_m))) +vfloat64m4_t vfsgnj(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m8))) +vfloat64m8_t vfsgnj(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m8_m))) +vfloat64m8_t vfsgnj(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m1))) +vfloat64m1_t vfsgnjn(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m1_m))) +vfloat64m1_t vfsgnjn(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m2))) +vfloat64m2_t vfsgnjn(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m2_m))) +vfloat64m2_t vfsgnjn(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m4))) +vfloat64m4_t vfsgnjn(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m4_m))) +vfloat64m4_t vfsgnjn(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m8))) +vfloat64m8_t vfsgnjn(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m8_m))) +vfloat64m8_t vfsgnjn(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m1))) +vfloat64m1_t vfsgnjn(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m1_m))) +vfloat64m1_t vfsgnjn(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m2))) +vfloat64m2_t vfsgnjn(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m2_m))) +vfloat64m2_t vfsgnjn(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m4))) +vfloat64m4_t vfsgnjn(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m4_m))) +vfloat64m4_t vfsgnjn(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m8))) +vfloat64m8_t vfsgnjn(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m8_m))) +vfloat64m8_t vfsgnjn(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m1))) +vfloat64m1_t vfsgnjx(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m1_m))) +vfloat64m1_t vfsgnjx(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m2))) +vfloat64m2_t vfsgnjx(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m2_m))) +vfloat64m2_t vfsgnjx(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m4))) +vfloat64m4_t vfsgnjx(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m4_m))) +vfloat64m4_t vfsgnjx(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m8))) +vfloat64m8_t vfsgnjx(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m8_m))) +vfloat64m8_t vfsgnjx(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m1))) +vfloat64m1_t vfsgnjx(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m1_m))) +vfloat64m1_t vfsgnjx(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m2))) +vfloat64m2_t vfsgnjx(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m2_m))) +vfloat64m2_t vfsgnjx(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m4))) +vfloat64m4_t vfsgnjx(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m4_m))) +vfloat64m4_t vfsgnjx(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m8))) +vfloat64m8_t vfsgnjx(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m8_m))) +vfloat64m8_t vfsgnjx(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m1))) +vfloat64m1_t vfabs(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m1_m))) +vfloat64m1_t vfabs(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m2))) +vfloat64m2_t vfabs(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m2_m))) +vfloat64m2_t vfabs(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m4))) +vfloat64m4_t vfabs(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m4_m))) +vfloat64m4_t vfabs(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m8))) +vfloat64m8_t vfabs(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m8_m))) +vfloat64m8_t vfabs(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m1_b64))) +vbool64_t vmfeq(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m1_b64_m))) +vbool64_t vmfeq(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m2_b32))) +vbool32_t vmfeq(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m2_b32_m))) +vbool32_t vmfeq(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m4_b16))) +vbool16_t vmfeq(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m4_b16_m))) +vbool16_t vmfeq(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m8_b8))) +vbool8_t vmfeq(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m8_b8_m))) +vbool8_t vmfeq(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m1_b64))) +vbool64_t vmfeq(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m1_b64_m))) +vbool64_t vmfeq(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m2_b32))) +vbool32_t vmfeq(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m2_b32_m))) +vbool32_t vmfeq(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m4_b16))) +vbool16_t vmfeq(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m4_b16_m))) +vbool16_t vmfeq(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m8_b8))) +vbool8_t vmfeq(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m8_b8_m))) +vbool8_t vmfeq(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m1_b64))) +vbool64_t vmfne(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m1_b64_m))) +vbool64_t vmfne(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m2_b32))) +vbool32_t vmfne(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m2_b32_m))) +vbool32_t vmfne(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m4_b16))) +vbool16_t vmfne(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m4_b16_m))) +vbool16_t vmfne(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m8_b8))) +vbool8_t vmfne(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m8_b8_m))) +vbool8_t vmfne(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m1_b64))) +vbool64_t vmfne(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m1_b64_m))) +vbool64_t vmfne(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m2_b32))) +vbool32_t vmfne(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m2_b32_m))) +vbool32_t vmfne(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m4_b16))) +vbool16_t vmfne(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m4_b16_m))) +vbool16_t vmfne(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m8_b8))) +vbool8_t vmfne(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m8_b8_m))) +vbool8_t vmfne(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m1_b64))) +vbool64_t vmflt(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m1_b64_m))) +vbool64_t vmflt(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m2_b32))) +vbool32_t vmflt(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m2_b32_m))) +vbool32_t vmflt(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m4_b16))) +vbool16_t vmflt(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m4_b16_m))) +vbool16_t vmflt(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m8_b8))) +vbool8_t vmflt(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m8_b8_m))) +vbool8_t vmflt(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m1_b64))) +vbool64_t vmflt(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m1_b64_m))) +vbool64_t vmflt(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m2_b32))) +vbool32_t vmflt(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m2_b32_m))) +vbool32_t vmflt(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m4_b16))) +vbool16_t vmflt(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m4_b16_m))) +vbool16_t vmflt(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m8_b8))) +vbool8_t vmflt(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m8_b8_m))) +vbool8_t vmflt(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m1_b64))) +vbool64_t vmfle(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m1_b64_m))) +vbool64_t vmfle(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m2_b32))) +vbool32_t vmfle(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m2_b32_m))) +vbool32_t vmfle(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m4_b16))) +vbool16_t vmfle(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m4_b16_m))) +vbool16_t vmfle(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m8_b8))) +vbool8_t vmfle(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m8_b8_m))) +vbool8_t vmfle(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m1_b64))) +vbool64_t vmfle(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m1_b64_m))) +vbool64_t vmfle(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m2_b32))) +vbool32_t vmfle(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m2_b32_m))) +vbool32_t vmfle(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m4_b16))) +vbool16_t vmfle(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m4_b16_m))) +vbool16_t vmfle(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m8_b8))) +vbool8_t vmfle(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m8_b8_m))) +vbool8_t vmfle(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m1_b64))) +vbool64_t vmfgt(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m1_b64_m))) +vbool64_t vmfgt(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m2_b32))) +vbool32_t vmfgt(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m2_b32_m))) +vbool32_t vmfgt(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m4_b16))) +vbool16_t vmfgt(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m4_b16_m))) +vbool16_t vmfgt(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m8_b8))) +vbool8_t vmfgt(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m8_b8_m))) +vbool8_t vmfgt(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m1_b64))) +vbool64_t vmfgt(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m1_b64_m))) +vbool64_t vmfgt(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m2_b32))) +vbool32_t vmfgt(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m2_b32_m))) +vbool32_t vmfgt(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m4_b16))) +vbool16_t vmfgt(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m4_b16_m))) +vbool16_t vmfgt(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m8_b8))) +vbool8_t vmfgt(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m8_b8_m))) +vbool8_t vmfgt(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m1_b64))) +vbool64_t vmfge(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m1_b64_m))) +vbool64_t vmfge(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m2_b32))) +vbool32_t vmfge(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m2_b32_m))) +vbool32_t vmfge(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m4_b16))) +vbool16_t vmfge(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m4_b16_m))) +vbool16_t vmfge(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m8_b8))) +vbool8_t vmfge(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m8_b8_m))) +vbool8_t vmfge(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m1_b64))) +vbool64_t vmfge(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m1_b64_m))) +vbool64_t vmfge(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m2_b32))) +vbool32_t vmfge(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m2_b32_m))) +vbool32_t vmfge(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m4_b16))) +vbool16_t vmfge(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m4_b16_m))) +vbool16_t vmfge(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m8_b8))) +vbool8_t vmfge(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m8_b8_m))) +vbool8_t vmfge(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f64m1))) +vfloat64m1_t vmerge(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f64m2))) +vfloat64m2_t vmerge(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f64m4))) +vfloat64m4_t vmerge(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f64m8))) +vfloat64m8_t vmerge(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f64m1))) +vfloat64m1_t vfmerge(vbool64_t op0, vfloat64m1_t op1, double op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f64m2))) +vfloat64m2_t vfmerge(vbool32_t op0, vfloat64m2_t op1, double op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f64m4))) +vfloat64m4_t vfmerge(vbool16_t op0, vfloat64m4_t op1, double op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f64m8))) +vfloat64m8_t vfmerge(vbool8_t op0, vfloat64m8_t op1, double op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m1_f64m1))) +vfloat64m1_t vfredmax(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m1_f64m1_m))) +vfloat64m1_t vfredmax(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m2_f64m1))) +vfloat64m1_t vfredmax(vfloat64m1_t op0, vfloat64m2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m2_f64m1_m))) +vfloat64m1_t vfredmax(vbool32_t op0, vfloat64m1_t op1, vfloat64m2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m4_f64m1))) +vfloat64m1_t vfredmax(vfloat64m1_t op0, vfloat64m4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m4_f64m1_m))) +vfloat64m1_t vfredmax(vbool16_t op0, vfloat64m1_t op1, vfloat64m4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m8_f64m1))) +vfloat64m1_t vfredmax(vfloat64m1_t op0, vfloat64m8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m8_f64m1_m))) +vfloat64m1_t vfredmax(vbool8_t op0, vfloat64m1_t op1, vfloat64m8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m1_f64m1))) +vfloat64m1_t vfredmin(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m1_f64m1_m))) +vfloat64m1_t vfredmin(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m2_f64m1))) +vfloat64m1_t vfredmin(vfloat64m1_t op0, vfloat64m2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m2_f64m1_m))) +vfloat64m1_t vfredmin(vbool32_t op0, vfloat64m1_t op1, vfloat64m2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m4_f64m1))) +vfloat64m1_t vfredmin(vfloat64m1_t op0, vfloat64m4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m4_f64m1_m))) +vfloat64m1_t vfredmin(vbool16_t op0, vfloat64m1_t op1, vfloat64m4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m8_f64m1))) +vfloat64m1_t vfredmin(vfloat64m1_t op0, vfloat64m8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m8_f64m1_m))) +vfloat64m1_t vfredmin(vbool8_t op0, vfloat64m1_t op1, vfloat64m8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m1_f64m1))) +vfloat64m1_t vfredsum(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m1_f64m1_m))) +vfloat64m1_t vfredsum(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m2_f64m1))) +vfloat64m1_t vfredsum(vfloat64m1_t op0, vfloat64m2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m2_f64m1_m))) +vfloat64m1_t vfredsum(vbool32_t op0, vfloat64m1_t op1, vfloat64m2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m4_f64m1))) +vfloat64m1_t vfredsum(vfloat64m1_t op0, vfloat64m4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m4_f64m1_m))) +vfloat64m1_t vfredsum(vbool16_t op0, vfloat64m1_t op1, vfloat64m4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m8_f64m1))) +vfloat64m1_t vfredsum(vfloat64m1_t op0, vfloat64m8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m8_f64m1_m))) +vfloat64m1_t vfredsum(vbool8_t op0, vfloat64m1_t op1, vfloat64m8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m1_f64m1))) +vfloat64m1_t vfredosum(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m1_f64m1_m))) +vfloat64m1_t vfredosum(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m2_f64m1))) +vfloat64m1_t vfredosum(vfloat64m1_t op0, vfloat64m2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m2_f64m1_m))) +vfloat64m1_t vfredosum(vbool32_t op0, vfloat64m1_t op1, vfloat64m2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m4_f64m1))) +vfloat64m1_t vfredosum(vfloat64m1_t op0, vfloat64m4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m4_f64m1_m))) +vfloat64m1_t vfredosum(vbool16_t op0, vfloat64m1_t op1, vfloat64m4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m8_f64m1))) +vfloat64m1_t vfredosum(vfloat64m1_t op0, vfloat64m8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m8_f64m1_m))) +vfloat64m1_t vfredosum(vbool8_t op0, vfloat64m1_t op1, vfloat64m8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f64m1_f64))) +double vfmv_f(vfloat64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f64m2_f64))) +double vfmv_f(vfloat64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f64m4_f64))) +double vfmv_f(vfloat64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f64m8_f64))) +double vfmv_f(vfloat64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f64m1))) +vfloat64m1_t vfmv_s(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f64m2))) +vfloat64m2_t vfmv_s(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f64m4))) +vfloat64m4_t vfmv_s(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f64m8))) +vfloat64m8_t vfmv_s(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m1))) +vfloat64m1_t vslideup(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m1_m))) +vfloat64m1_t vslideup(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m2))) +vfloat64m2_t vslideup(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m2_m))) +vfloat64m2_t vslideup(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m4))) +vfloat64m4_t vslideup(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m4_m))) +vfloat64m4_t vslideup(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m8))) +vfloat64m8_t vslideup(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m8_m))) +vfloat64m8_t vslideup(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m1))) +vfloat64m1_t vslidedown(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m1_m))) +vfloat64m1_t vslidedown(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m2))) +vfloat64m2_t vslidedown(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m2_m))) +vfloat64m2_t vslidedown(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m4))) +vfloat64m4_t vslidedown(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m4_m))) +vfloat64m4_t vslidedown(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m8))) +vfloat64m8_t vslidedown(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m8_m))) +vfloat64m8_t vslidedown(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m1))) +vfloat64m1_t vfslide1up(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m1_m))) +vfloat64m1_t vfslide1up(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m2))) +vfloat64m2_t vfslide1up(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m2_m))) +vfloat64m2_t vfslide1up(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m4))) +vfloat64m4_t vfslide1up(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m4_m))) +vfloat64m4_t vfslide1up(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m8))) +vfloat64m8_t vfslide1up(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m8_m))) +vfloat64m8_t vfslide1up(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m1))) +vfloat64m1_t vfslide1down(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m1_m))) +vfloat64m1_t vfslide1down(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m2))) +vfloat64m2_t vfslide1down(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m2_m))) +vfloat64m2_t vfslide1down(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m4))) +vfloat64m4_t vfslide1down(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m4_m))) +vfloat64m4_t vfslide1down(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m8))) +vfloat64m8_t vfslide1down(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m8_m))) +vfloat64m8_t vfslide1down(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m1))) +vfloat64m1_t vrgather(vfloat64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m1_m))) +vfloat64m1_t vrgather(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m2))) +vfloat64m2_t vrgather(vfloat64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m2_m))) +vfloat64m2_t vrgather(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m4))) +vfloat64m4_t vrgather(vfloat64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m4_m))) +vfloat64m4_t vrgather(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m8))) +vfloat64m8_t vrgather(vfloat64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m8_m))) +vfloat64m8_t vrgather(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m1))) +vfloat64m1_t vrgather(vfloat64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m1_m))) +vfloat64m1_t vrgather(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m2))) +vfloat64m2_t vrgather(vfloat64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m2_m))) +vfloat64m2_t vrgather(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m4))) +vfloat64m4_t vrgather(vfloat64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m4_m))) +vfloat64m4_t vrgather(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m8))) +vfloat64m8_t vrgather(vfloat64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m8_m))) +vfloat64m8_t vrgather(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m1))) +vfloat64m1_t vrgatherei16(vfloat64m1_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m1_m))) +vfloat64m1_t vrgatherei16(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m2))) +vfloat64m2_t vrgatherei16(vfloat64m2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m2_m))) +vfloat64m2_t vrgatherei16(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m4))) +vfloat64m4_t vrgatherei16(vfloat64m4_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m4_m))) +vfloat64m4_t vrgatherei16(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m8))) +vfloat64m8_t vrgatherei16(vfloat64m8_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m8_m))) +vfloat64m8_t vrgatherei16(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f64m1))) +vfloat64m1_t vcompress(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f64m2))) +vfloat64m2_t vcompress(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f64m4))) +vfloat64m4_t vcompress(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f64m8))) +vfloat64m8_t vcompress(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f64m2_f64m1))) +vfloat64m1_t vget_f64m1(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f64m4_f64m1))) +vfloat64m1_t vget_f64m1(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f64m8_f64m1))) +vfloat64m1_t vget_f64m1(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f64m4_f64m2))) +vfloat64m2_t vget_f64m2(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f64m8_f64m2))) +vfloat64m2_t vget_f64m2(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f64m8_f64m4))) +vfloat64m4_t vget_f64m4(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f64m1_f64m2))) +vfloat64m2_t vset(vfloat64m2_t op0, size_t op1, vfloat64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f64m1_f64m4))) +vfloat64m4_t vset(vfloat64m4_t op0, size_t op1, vfloat64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f64m2_f64m4))) +vfloat64m4_t vset(vfloat64m4_t op0, size_t op1, vfloat64m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f64m1_f64m8))) +vfloat64m8_t vset(vfloat64m8_t op0, size_t op1, vfloat64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f64m2_f64m8))) +vfloat64m8_t vset(vfloat64m8_t op0, size_t op1, vfloat64m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f64m4_f64m8))) +vfloat64m8_t vset(vfloat64m8_t op0, size_t op1, vfloat64m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m1))) +void vsuxei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m1_m))) +void vsuxei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m2))) +void vsuxei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m2_m))) +void vsuxei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m4))) +void vsuxei8(double * op0, vuint8mf2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m4_m))) +void vsuxei8(vbool16_t op0, double * op1, vuint8mf2_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m8))) +void vsuxei8(double * op0, vuint8m1_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m8_m))) +void vsuxei8(vbool8_t op0, double * op1, vuint8m1_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m1))) +void vsuxei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m1_m))) +void vsuxei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m2))) +void vsuxei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m2_m))) +void vsuxei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m4))) +void vsuxei16(double * op0, vuint16m1_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m4_m))) +void vsuxei16(vbool16_t op0, double * op1, vuint16m1_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m8))) +void vsuxei16(double * op0, vuint16m2_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m8_m))) +void vsuxei16(vbool8_t op0, double * op1, vuint16m2_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m1))) +void vsuxei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m1_m))) +void vsuxei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m2))) +void vsuxei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m2_m))) +void vsuxei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m4))) +void vsuxei32(double * op0, vuint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m4_m))) +void vsuxei32(vbool16_t op0, double * op1, vuint32m2_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m8))) +void vsuxei32(double * op0, vuint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m8_m))) +void vsuxei32(vbool8_t op0, double * op1, vuint32m4_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m1))) +void vsuxei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m1_m))) +void vsuxei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m2))) +void vsuxei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m2_m))) +void vsuxei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m4))) +void vsuxei64(double * op0, vuint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m4_m))) +void vsuxei64(vbool16_t op0, double * op1, vuint64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m8))) +void vsuxei64(double * op0, vuint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m8_m))) +void vsuxei64(vbool8_t op0, double * op1, vuint64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m1))) +void vsoxei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m1_m))) +void vsoxei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m2))) +void vsoxei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m2_m))) +void vsoxei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m4))) +void vsoxei8(double * op0, vuint8mf2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m4_m))) +void vsoxei8(vbool16_t op0, double * op1, vuint8mf2_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m8))) +void vsoxei8(double * op0, vuint8m1_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m8_m))) +void vsoxei8(vbool8_t op0, double * op1, vuint8m1_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m1))) +void vsoxei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m1_m))) +void vsoxei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m2))) +void vsoxei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m2_m))) +void vsoxei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m4))) +void vsoxei16(double * op0, vuint16m1_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m4_m))) +void vsoxei16(vbool16_t op0, double * op1, vuint16m1_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m8))) +void vsoxei16(double * op0, vuint16m2_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m8_m))) +void vsoxei16(vbool8_t op0, double * op1, vuint16m2_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m1))) +void vsoxei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m1_m))) +void vsoxei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m2))) +void vsoxei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m2_m))) +void vsoxei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m4))) +void vsoxei32(double * op0, vuint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m4_m))) +void vsoxei32(vbool16_t op0, double * op1, vuint32m2_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m8))) +void vsoxei32(double * op0, vuint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m8_m))) +void vsoxei32(vbool8_t op0, double * op1, vuint32m4_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m1))) +void vsoxei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m1_m))) +void vsoxei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m2))) +void vsoxei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m2_m))) +void vsoxei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m4))) +void vsoxei64(double * op0, vuint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m4_m))) +void vsoxei64(vbool16_t op0, double * op1, vuint64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m8))) +void vsoxei64(double * op0, vuint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m8_m))) +void vsoxei64(vbool8_t op0, double * op1, vuint64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_f64m1_m))) +vfloat64m1_t vle64ff(vbool64_t op0, vfloat64m1_t op1, const double * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_f64m2_m))) +vfloat64m2_t vle64ff(vbool32_t op0, vfloat64m2_t op1, const double * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_f64m4_m))) +vfloat64m4_t vle64ff(vbool16_t op0, vfloat64m4_t op1, const double * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_f64m8_m))) +vfloat64m8_t vle64ff(vbool8_t op0, vfloat64m8_t op1, const double * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m1))) +vfloat64m1_t vfneg(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m1_m))) +vfloat64m1_t vfneg(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m2))) +vfloat64m2_t vfneg(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m2_m))) +vfloat64m2_t vfneg(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m4))) +vfloat64m4_t vfneg(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m4_m))) +vfloat64m4_t vfneg(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m8))) +vfloat64m8_t vfneg(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m8_m))) +vfloat64m8_t vfneg(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_f64m1_m))) +vfloat64m1_t vle64(vbool64_t op0, vfloat64m1_t op1, const double * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_f64m2_m))) +vfloat64m2_t vle64(vbool32_t op0, vfloat64m2_t op1, const double * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_f64m4_m))) +vfloat64m4_t vle64(vbool16_t op0, vfloat64m4_t op1, const double * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_f64m8_m))) +vfloat64m8_t vle64(vbool8_t op0, vfloat64m8_t op1, const double * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m1))) +void vse64(double * op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m1_m))) +void vse64(vbool64_t op0, double * op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m2))) +void vse64(double * op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m2_m))) +void vse64(vbool32_t op0, double * op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m4))) +void vse64(double * op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m4_m))) +void vse64(vbool16_t op0, double * op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m8))) +void vse64(double * op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m8_m))) +void vse64(vbool8_t op0, double * op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m1))) +vuint64m1_t vfclass(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m1_m))) +vuint64m1_t vfclass(vbool64_t op0, vuint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m2))) +vuint64m2_t vfclass(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m2_m))) +vuint64m2_t vfclass(vbool32_t op0, vuint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m4))) +vuint64m4_t vfclass(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m4_m))) +vuint64m4_t vfclass(vbool16_t op0, vuint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m8))) +vuint64m8_t vfclass(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m8_m))) +vuint64m8_t vfclass(vbool8_t op0, vuint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m1))) +vfloat64m1_t vfcvt_f(vint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m1_m))) +vfloat64m1_t vfcvt_f(vbool64_t op0, vfloat64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m2))) +vfloat64m2_t vfcvt_f(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m2_m))) +vfloat64m2_t vfcvt_f(vbool32_t op0, vfloat64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m4))) +vfloat64m4_t vfcvt_f(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m4_m))) +vfloat64m4_t vfcvt_f(vbool16_t op0, vfloat64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m8))) +vfloat64m8_t vfcvt_f(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m8_m))) +vfloat64m8_t vfcvt_f(vbool8_t op0, vfloat64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m1))) +vfloat64m1_t vfcvt_f(vuint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m1_m))) +vfloat64m1_t vfcvt_f(vbool64_t op0, vfloat64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m2))) +vfloat64m2_t vfcvt_f(vuint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m2_m))) +vfloat64m2_t vfcvt_f(vbool32_t op0, vfloat64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m4))) +vfloat64m4_t vfcvt_f(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m4_m))) +vfloat64m4_t vfcvt_f(vbool16_t op0, vfloat64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m8))) +vfloat64m8_t vfcvt_f(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m8_m))) +vfloat64m8_t vfcvt_f(vbool8_t op0, vfloat64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m1))) +vint64m1_t vfcvt_rtz_x(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m1_m))) +vint64m1_t vfcvt_rtz_x(vbool64_t op0, vint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m2))) +vint64m2_t vfcvt_rtz_x(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m2_m))) +vint64m2_t vfcvt_rtz_x(vbool32_t op0, vint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m4))) +vint64m4_t vfcvt_rtz_x(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m4_m))) +vint64m4_t vfcvt_rtz_x(vbool16_t op0, vint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m8))) +vint64m8_t vfcvt_rtz_x(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m8_m))) +vint64m8_t vfcvt_rtz_x(vbool8_t op0, vint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m1))) +vuint64m1_t vfcvt_rtz_xu(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m1_m))) +vuint64m1_t vfcvt_rtz_xu(vbool64_t op0, vuint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m2))) +vuint64m2_t vfcvt_rtz_xu(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m2_m))) +vuint64m2_t vfcvt_rtz_xu(vbool32_t op0, vuint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m4))) +vuint64m4_t vfcvt_rtz_xu(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m4_m))) +vuint64m4_t vfcvt_rtz_xu(vbool16_t op0, vuint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m8))) +vuint64m8_t vfcvt_rtz_xu(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m8_m))) +vuint64m8_t vfcvt_rtz_xu(vbool8_t op0, vuint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m1))) +vint64m1_t vfcvt_x(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m1_m))) +vint64m1_t vfcvt_x(vbool64_t op0, vint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m2))) +vint64m2_t vfcvt_x(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m2_m))) +vint64m2_t vfcvt_x(vbool32_t op0, vint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m4))) +vint64m4_t vfcvt_x(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m4_m))) +vint64m4_t vfcvt_x(vbool16_t op0, vint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m8))) +vint64m8_t vfcvt_x(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m8_m))) +vint64m8_t vfcvt_x(vbool8_t op0, vint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m1))) +vuint64m1_t vfcvt_xu(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m1_m))) +vuint64m1_t vfcvt_xu(vbool64_t op0, vuint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m2))) +vuint64m2_t vfcvt_xu(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m2_m))) +vuint64m2_t vfcvt_xu(vbool32_t op0, vuint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m4))) +vuint64m4_t vfcvt_xu(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m4_m))) +vuint64m4_t vfcvt_xu(vbool16_t op0, vuint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m8))) +vuint64m8_t vfcvt_xu(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m8_m))) +vuint64m8_t vfcvt_xu(vbool8_t op0, vuint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32mf2))) +vint32mf2_t vfncvt_rtz_x(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32mf2_m))) +vint32mf2_t vfncvt_rtz_x(vbool64_t op0, vint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32m1))) +vint32m1_t vfncvt_rtz_x(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32m1_m))) +vint32m1_t vfncvt_rtz_x(vbool32_t op0, vint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32m2))) +vint32m2_t vfncvt_rtz_x(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32m2_m))) +vint32m2_t vfncvt_rtz_x(vbool16_t op0, vint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32m4))) +vint32m4_t vfncvt_rtz_x(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32m4_m))) +vint32m4_t vfncvt_rtz_x(vbool8_t op0, vint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32mf2))) +vuint32mf2_t vfncvt_rtz_xu(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32mf2_m))) +vuint32mf2_t vfncvt_rtz_xu(vbool64_t op0, vuint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32m1))) +vuint32m1_t vfncvt_rtz_xu(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32m1_m))) +vuint32m1_t vfncvt_rtz_xu(vbool32_t op0, vuint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32m2))) +vuint32m2_t vfncvt_rtz_xu(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32m2_m))) +vuint32m2_t vfncvt_rtz_xu(vbool16_t op0, vuint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32m4))) +vuint32m4_t vfncvt_rtz_xu(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32m4_m))) +vuint32m4_t vfncvt_rtz_xu(vbool8_t op0, vuint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32mf2))) +vint32mf2_t vfncvt_x(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32mf2_m))) +vint32mf2_t vfncvt_x(vbool64_t op0, vint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32m1))) +vint32m1_t vfncvt_x(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32m1_m))) +vint32m1_t vfncvt_x(vbool32_t op0, vint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32m2))) +vint32m2_t vfncvt_x(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32m2_m))) +vint32m2_t vfncvt_x(vbool16_t op0, vint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32m4))) +vint32m4_t vfncvt_x(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32m4_m))) +vint32m4_t vfncvt_x(vbool8_t op0, vint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32mf2))) +vuint32mf2_t vfncvt_xu(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32mf2_m))) +vuint32mf2_t vfncvt_xu(vbool64_t op0, vuint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32m1))) +vuint32m1_t vfncvt_xu(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32m1_m))) +vuint32m1_t vfncvt_xu(vbool32_t op0, vuint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32m2))) +vuint32m2_t vfncvt_xu(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32m2_m))) +vuint32m2_t vfncvt_xu(vbool16_t op0, vuint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32m4))) +vuint32m4_t vfncvt_xu(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32m4_m))) +vuint32m4_t vfncvt_xu(vbool8_t op0, vuint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m1))) +vfloat64m1_t vfrec7(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m1_m))) +vfloat64m1_t vfrec7(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m2))) +vfloat64m2_t vfrec7(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m2_m))) +vfloat64m2_t vfrec7(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m4))) +vfloat64m4_t vfrec7(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m4_m))) +vfloat64m4_t vfrec7(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m8))) +vfloat64m8_t vfrec7(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m8_m))) +vfloat64m8_t vfrec7(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m1))) +vfloat64m1_t vfrsqrt7(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m1_m))) +vfloat64m1_t vfrsqrt7(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m2))) +vfloat64m2_t vfrsqrt7(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m2_m))) +vfloat64m2_t vfrsqrt7(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m4))) +vfloat64m4_t vfrsqrt7(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m4_m))) +vfloat64m4_t vfrsqrt7(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m8))) +vfloat64m8_t vfrsqrt7(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m8_m))) +vfloat64m8_t vfrsqrt7(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m1))) +vfloat64m1_t vfsqrt(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m1_m))) +vfloat64m1_t vfsqrt(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m2))) +vfloat64m2_t vfsqrt(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m2_m))) +vfloat64m2_t vfsqrt(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m4))) +vfloat64m4_t vfsqrt(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m4_m))) +vfloat64m4_t vfsqrt(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m8))) +vfloat64m8_t vfsqrt(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m8_m))) +vfloat64m8_t vfsqrt(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m1))) +vfloat64m1_t vfwcvt_f(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m1_m))) +vfloat64m1_t vfwcvt_f(vbool64_t op0, vfloat64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m2))) +vfloat64m2_t vfwcvt_f(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m2_m))) +vfloat64m2_t vfwcvt_f(vbool32_t op0, vfloat64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m4))) +vfloat64m4_t vfwcvt_f(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m4_m))) +vfloat64m4_t vfwcvt_f(vbool16_t op0, vfloat64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m8))) +vfloat64m8_t vfwcvt_f(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m8_m))) +vfloat64m8_t vfwcvt_f(vbool8_t op0, vfloat64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m1))) +vfloat64m1_t vfwcvt_f(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m1_m))) +vfloat64m1_t vfwcvt_f(vbool64_t op0, vfloat64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m2))) +vfloat64m2_t vfwcvt_f(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m2_m))) +vfloat64m2_t vfwcvt_f(vbool32_t op0, vfloat64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m4))) +vfloat64m4_t vfwcvt_f(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m4_m))) +vfloat64m4_t vfwcvt_f(vbool16_t op0, vfloat64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m8))) +vfloat64m8_t vfwcvt_f(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m8_m))) +vfloat64m8_t vfwcvt_f(vbool8_t op0, vfloat64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f64m1_f64m2))) +vfloat64m2_t vlmul_ext_f64m2(vfloat64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f64m1_f64m4))) +vfloat64m4_t vlmul_ext_f64m4(vfloat64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f64m2_f64m4))) +vfloat64m4_t vlmul_ext_f64m4(vfloat64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f64m1_f64m8))) +vfloat64m8_t vlmul_ext_f64m8(vfloat64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f64m2_f64m8))) +vfloat64m8_t vlmul_ext_f64m8(vfloat64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f64m4_f64m8))) +vfloat64m8_t vlmul_ext_f64m8(vfloat64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f64m2_f64m1))) +vfloat64m1_t vlmul_trunc_f64m1(vfloat64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f64m4_f64m1))) +vfloat64m1_t vlmul_trunc_f64m1(vfloat64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f64m8_f64m1))) +vfloat64m1_t vlmul_trunc_f64m1(vfloat64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f64m4_f64m2))) +vfloat64m2_t vlmul_trunc_f64m2(vfloat64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f64m8_f64m2))) +vfloat64m2_t vlmul_trunc_f64m2(vfloat64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f64m8_f64m4))) +vfloat64m4_t vlmul_trunc_f64m4(vfloat64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m1_f64m1))) +vfloat64m1_t vreinterpret_f64m1(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m2_f64m2))) +vfloat64m2_t vreinterpret_f64m2(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m4_f64m4))) +vfloat64m4_t vreinterpret_f64m4(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m8_f64m8))) +vfloat64m8_t vreinterpret_f64m8(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m1_f64m1))) +vfloat64m1_t vreinterpret_f64m1(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m2_f64m2))) +vfloat64m2_t vreinterpret_f64m2(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m4_f64m4))) +vfloat64m4_t vreinterpret_f64m4(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m8_f64m8))) +vfloat64m8_t vreinterpret_f64m8(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m1_i64m1))) +vint64m1_t vreinterpret_i64m1(vfloat64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m2_i64m2))) +vint64m2_t vreinterpret_i64m2(vfloat64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m4_i64m4))) +vint64m4_t vreinterpret_i64m4(vfloat64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m8_i64m8))) +vint64m8_t vreinterpret_i64m8(vfloat64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m1_u64m1))) +vuint64m1_t vreinterpret_u64m1(vfloat64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m2_u64m2))) +vuint64m2_t vreinterpret_u64m2(vfloat64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m4_u64m4))) +vuint64m4_t vreinterpret_u64m4(vfloat64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m8_u64m8))) +vuint64m8_t vreinterpret_u64m8(vfloat64m8_t op0); + +#endif + +#if defined(__riscv_f) && defined(__riscv_d) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m1))) +vfloat64m1_t vfwadd_vv(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m1_m))) +vfloat64m1_t vfwadd_vv(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m2))) +vfloat64m2_t vfwadd_vv(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m2_m))) +vfloat64m2_t vfwadd_vv(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m4))) +vfloat64m4_t vfwadd_vv(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m4_m))) +vfloat64m4_t vfwadd_vv(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m8))) +vfloat64m8_t vfwadd_vv(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m8_m))) +vfloat64m8_t vfwadd_vv(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m1))) +vfloat64m1_t vfwadd_vf(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m1_m))) +vfloat64m1_t vfwadd_vf(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m2))) +vfloat64m2_t vfwadd_vf(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m2_m))) +vfloat64m2_t vfwadd_vf(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m4))) +vfloat64m4_t vfwadd_vf(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m4_m))) +vfloat64m4_t vfwadd_vf(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m8))) +vfloat64m8_t vfwadd_vf(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m8_m))) +vfloat64m8_t vfwadd_vf(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m1))) +vfloat64m1_t vfwsub_vv(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m1_m))) +vfloat64m1_t vfwsub_vv(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m2))) +vfloat64m2_t vfwsub_vv(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m2_m))) +vfloat64m2_t vfwsub_vv(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m4))) +vfloat64m4_t vfwsub_vv(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m4_m))) +vfloat64m4_t vfwsub_vv(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m8))) +vfloat64m8_t vfwsub_vv(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m8_m))) +vfloat64m8_t vfwsub_vv(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m1))) +vfloat64m1_t vfwsub_vf(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m1_m))) +vfloat64m1_t vfwsub_vf(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m2))) +vfloat64m2_t vfwsub_vf(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m2_m))) +vfloat64m2_t vfwsub_vf(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m4))) +vfloat64m4_t vfwsub_vf(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m4_m))) +vfloat64m4_t vfwsub_vf(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m8))) +vfloat64m8_t vfwsub_vf(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m8_m))) +vfloat64m8_t vfwsub_vf(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m1))) +vfloat64m1_t vfwadd_wv(vfloat64m1_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m1_m))) +vfloat64m1_t vfwadd_wv(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m2))) +vfloat64m2_t vfwadd_wv(vfloat64m2_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m2_m))) +vfloat64m2_t vfwadd_wv(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m4))) +vfloat64m4_t vfwadd_wv(vfloat64m4_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m4_m))) +vfloat64m4_t vfwadd_wv(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m8))) +vfloat64m8_t vfwadd_wv(vfloat64m8_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m8_m))) +vfloat64m8_t vfwadd_wv(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m1))) +vfloat64m1_t vfwadd_wf(vfloat64m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m1_m))) +vfloat64m1_t vfwadd_wf(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m2))) +vfloat64m2_t vfwadd_wf(vfloat64m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m2_m))) +vfloat64m2_t vfwadd_wf(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m4))) +vfloat64m4_t vfwadd_wf(vfloat64m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m4_m))) +vfloat64m4_t vfwadd_wf(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m8))) +vfloat64m8_t vfwadd_wf(vfloat64m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m8_m))) +vfloat64m8_t vfwadd_wf(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m1))) +vfloat64m1_t vfwsub_wv(vfloat64m1_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m1_m))) +vfloat64m1_t vfwsub_wv(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m2))) +vfloat64m2_t vfwsub_wv(vfloat64m2_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m2_m))) +vfloat64m2_t vfwsub_wv(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m4))) +vfloat64m4_t vfwsub_wv(vfloat64m4_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m4_m))) +vfloat64m4_t vfwsub_wv(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m8))) +vfloat64m8_t vfwsub_wv(vfloat64m8_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m8_m))) +vfloat64m8_t vfwsub_wv(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m1))) +vfloat64m1_t vfwsub_wf(vfloat64m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m1_m))) +vfloat64m1_t vfwsub_wf(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m2))) +vfloat64m2_t vfwsub_wf(vfloat64m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m2_m))) +vfloat64m2_t vfwsub_wf(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m4))) +vfloat64m4_t vfwsub_wf(vfloat64m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m4_m))) +vfloat64m4_t vfwsub_wf(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m8))) +vfloat64m8_t vfwsub_wf(vfloat64m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m8_m))) +vfloat64m8_t vfwsub_wf(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m1))) +vfloat64m1_t vfwmul(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m1_m))) +vfloat64m1_t vfwmul(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m2))) +vfloat64m2_t vfwmul(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m2_m))) +vfloat64m2_t vfwmul(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m4))) +vfloat64m4_t vfwmul(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m4_m))) +vfloat64m4_t vfwmul(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m8))) +vfloat64m8_t vfwmul(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m8_m))) +vfloat64m8_t vfwmul(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m1))) +vfloat64m1_t vfwmul(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m1_m))) +vfloat64m1_t vfwmul(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m2))) +vfloat64m2_t vfwmul(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m2_m))) +vfloat64m2_t vfwmul(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m4))) +vfloat64m4_t vfwmul(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m4_m))) +vfloat64m4_t vfwmul(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m8))) +vfloat64m8_t vfwmul(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m8_m))) +vfloat64m8_t vfwmul(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m1))) +vfloat64m1_t vfwmacc(vfloat64m1_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m1_m))) +vfloat64m1_t vfwmacc(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m2))) +vfloat64m2_t vfwmacc(vfloat64m2_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m2_m))) +vfloat64m2_t vfwmacc(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m4))) +vfloat64m4_t vfwmacc(vfloat64m4_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m4_m))) +vfloat64m4_t vfwmacc(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m8))) +vfloat64m8_t vfwmacc(vfloat64m8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m8_m))) +vfloat64m8_t vfwmacc(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m1))) +vfloat64m1_t vfwmacc(vfloat64m1_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m1_m))) +vfloat64m1_t vfwmacc(vbool64_t op0, vfloat64m1_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m2))) +vfloat64m2_t vfwmacc(vfloat64m2_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m2_m))) +vfloat64m2_t vfwmacc(vbool32_t op0, vfloat64m2_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m4))) +vfloat64m4_t vfwmacc(vfloat64m4_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m4_m))) +vfloat64m4_t vfwmacc(vbool16_t op0, vfloat64m4_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m8))) +vfloat64m8_t vfwmacc(vfloat64m8_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m8_m))) +vfloat64m8_t vfwmacc(vbool8_t op0, vfloat64m8_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m1))) +vfloat64m1_t vfwnmacc(vfloat64m1_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m1_m))) +vfloat64m1_t vfwnmacc(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m2))) +vfloat64m2_t vfwnmacc(vfloat64m2_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m2_m))) +vfloat64m2_t vfwnmacc(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m4))) +vfloat64m4_t vfwnmacc(vfloat64m4_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m4_m))) +vfloat64m4_t vfwnmacc(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m8))) +vfloat64m8_t vfwnmacc(vfloat64m8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m8_m))) +vfloat64m8_t vfwnmacc(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m1))) +vfloat64m1_t vfwnmacc(vfloat64m1_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m1_m))) +vfloat64m1_t vfwnmacc(vbool64_t op0, vfloat64m1_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m2))) +vfloat64m2_t vfwnmacc(vfloat64m2_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m2_m))) +vfloat64m2_t vfwnmacc(vbool32_t op0, vfloat64m2_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m4))) +vfloat64m4_t vfwnmacc(vfloat64m4_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m4_m))) +vfloat64m4_t vfwnmacc(vbool16_t op0, vfloat64m4_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m8))) +vfloat64m8_t vfwnmacc(vfloat64m8_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m8_m))) +vfloat64m8_t vfwnmacc(vbool8_t op0, vfloat64m8_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m1))) +vfloat64m1_t vfwmsac(vfloat64m1_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m1_m))) +vfloat64m1_t vfwmsac(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m2))) +vfloat64m2_t vfwmsac(vfloat64m2_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m2_m))) +vfloat64m2_t vfwmsac(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m4))) +vfloat64m4_t vfwmsac(vfloat64m4_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m4_m))) +vfloat64m4_t vfwmsac(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m8))) +vfloat64m8_t vfwmsac(vfloat64m8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m8_m))) +vfloat64m8_t vfwmsac(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m1))) +vfloat64m1_t vfwmsac(vfloat64m1_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m1_m))) +vfloat64m1_t vfwmsac(vbool64_t op0, vfloat64m1_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m2))) +vfloat64m2_t vfwmsac(vfloat64m2_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m2_m))) +vfloat64m2_t vfwmsac(vbool32_t op0, vfloat64m2_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m4))) +vfloat64m4_t vfwmsac(vfloat64m4_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m4_m))) +vfloat64m4_t vfwmsac(vbool16_t op0, vfloat64m4_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m8))) +vfloat64m8_t vfwmsac(vfloat64m8_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m8_m))) +vfloat64m8_t vfwmsac(vbool8_t op0, vfloat64m8_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m1))) +vfloat64m1_t vfwnmsac(vfloat64m1_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m1_m))) +vfloat64m1_t vfwnmsac(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m2))) +vfloat64m2_t vfwnmsac(vfloat64m2_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m2_m))) +vfloat64m2_t vfwnmsac(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m4))) +vfloat64m4_t vfwnmsac(vfloat64m4_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m4_m))) +vfloat64m4_t vfwnmsac(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m8))) +vfloat64m8_t vfwnmsac(vfloat64m8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m8_m))) +vfloat64m8_t vfwnmsac(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m1))) +vfloat64m1_t vfwnmsac(vfloat64m1_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m1_m))) +vfloat64m1_t vfwnmsac(vbool64_t op0, vfloat64m1_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m2))) +vfloat64m2_t vfwnmsac(vfloat64m2_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m2_m))) +vfloat64m2_t vfwnmsac(vbool32_t op0, vfloat64m2_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m4))) +vfloat64m4_t vfwnmsac(vfloat64m4_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m4_m))) +vfloat64m4_t vfwnmsac(vbool16_t op0, vfloat64m4_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m8))) +vfloat64m8_t vfwnmsac(vfloat64m8_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m8_m))) +vfloat64m8_t vfwnmsac(vbool8_t op0, vfloat64m8_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m1_f64m1))) +vfloat64m1_t vfwredsum(vfloat64m1_t op0, vfloat32m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m1_f64m1_m))) +vfloat64m1_t vfwredsum(vbool32_t op0, vfloat64m1_t op1, vfloat32m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m2_f64m1))) +vfloat64m1_t vfwredsum(vfloat64m1_t op0, vfloat32m2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m2_f64m1_m))) +vfloat64m1_t vfwredsum(vbool16_t op0, vfloat64m1_t op1, vfloat32m2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m4_f64m1))) +vfloat64m1_t vfwredsum(vfloat64m1_t op0, vfloat32m4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m4_f64m1_m))) +vfloat64m1_t vfwredsum(vbool8_t op0, vfloat64m1_t op1, vfloat32m4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m8_f64m1))) +vfloat64m1_t vfwredsum(vfloat64m1_t op0, vfloat32m8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m8_f64m1_m))) +vfloat64m1_t vfwredsum(vbool4_t op0, vfloat64m1_t op1, vfloat32m8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32mf2_f64m1))) +vfloat64m1_t vfwredsum(vfloat64m1_t op0, vfloat32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32mf2_f64m1_m))) +vfloat64m1_t vfwredsum(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m1_f64m1))) +vfloat64m1_t vfwredosum(vfloat64m1_t op0, vfloat32m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m1_f64m1_m))) +vfloat64m1_t vfwredosum(vbool32_t op0, vfloat64m1_t op1, vfloat32m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m2_f64m1))) +vfloat64m1_t vfwredosum(vfloat64m1_t op0, vfloat32m2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m2_f64m1_m))) +vfloat64m1_t vfwredosum(vbool16_t op0, vfloat64m1_t op1, vfloat32m2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m4_f64m1))) +vfloat64m1_t vfwredosum(vfloat64m1_t op0, vfloat32m4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m4_f64m1_m))) +vfloat64m1_t vfwredosum(vbool8_t op0, vfloat64m1_t op1, vfloat32m4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m8_f64m1))) +vfloat64m1_t vfwredosum(vfloat64m1_t op0, vfloat32m8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m8_f64m1_m))) +vfloat64m1_t vfwredosum(vbool4_t op0, vfloat64m1_t op1, vfloat32m8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32mf2_f64m1))) +vfloat64m1_t vfwredosum(vfloat64m1_t op0, vfloat32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32mf2_f64m1_m))) +vfloat64m1_t vfwredosum(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32mf2))) +vfloat32mf2_t vfncvt_f(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32mf2_m))) +vfloat32mf2_t vfncvt_f(vbool64_t op0, vfloat32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32m1))) +vfloat32m1_t vfncvt_f(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32m1_m))) +vfloat32m1_t vfncvt_f(vbool32_t op0, vfloat32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32m2))) +vfloat32m2_t vfncvt_f(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32m2_m))) +vfloat32m2_t vfncvt_f(vbool16_t op0, vfloat32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32m4))) +vfloat32m4_t vfncvt_f(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32m4_m))) +vfloat32m4_t vfncvt_f(vbool8_t op0, vfloat32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32mf2))) +vfloat32mf2_t vfncvt_rod_f(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32mf2_m))) +vfloat32mf2_t vfncvt_rod_f(vbool64_t op0, vfloat32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32m1))) +vfloat32m1_t vfncvt_rod_f(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32m1_m))) +vfloat32m1_t vfncvt_rod_f(vbool32_t op0, vfloat32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32m2))) +vfloat32m2_t vfncvt_rod_f(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32m2_m))) +vfloat32m2_t vfncvt_rod_f(vbool16_t op0, vfloat32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32m4))) +vfloat32m4_t vfncvt_rod_f(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32m4_m))) +vfloat32m4_t vfncvt_rod_f(vbool8_t op0, vfloat32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m1))) +vfloat64m1_t vfwcvt_f(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m1_m))) +vfloat64m1_t vfwcvt_f(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m2))) +vfloat64m2_t vfwcvt_f(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m2_m))) +vfloat64m2_t vfwcvt_f(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m4))) +vfloat64m4_t vfwcvt_f(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m4_m))) +vfloat64m4_t vfwcvt_f(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m8))) +vfloat64m8_t vfwcvt_f(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m8_m))) +vfloat64m8_t vfwcvt_f(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, size_t op3); + +#endif + +#if defined(__riscv_zfh) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m1))) +void vsse16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m1_m))) +void vsse16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m2))) +void vsse16(_Float16 * op0, ptrdiff_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m2_m))) +void vsse16(vbool8_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m4))) +void vsse16(_Float16 * op0, ptrdiff_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m4_m))) +void vsse16(vbool4_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m8))) +void vsse16(_Float16 * op0, ptrdiff_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m8_m))) +void vsse16(vbool2_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16mf2))) +void vsse16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16mf2_m))) +void vsse16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16mf4))) +void vsse16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16mf4_m))) +void vsse16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m1))) +vfloat16m1_t vluxei8(const _Float16 * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m1_m))) +vfloat16m1_t vluxei8(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m2))) +vfloat16m2_t vluxei8(const _Float16 * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m2_m))) +vfloat16m2_t vluxei8(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m4))) +vfloat16m4_t vluxei8(const _Float16 * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m4_m))) +vfloat16m4_t vluxei8(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m8))) +vfloat16m8_t vluxei8(const _Float16 * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m8_m))) +vfloat16m8_t vluxei8(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16mf2))) +vfloat16mf2_t vluxei8(const _Float16 * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16mf2_m))) +vfloat16mf2_t vluxei8(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16mf4))) +vfloat16mf4_t vluxei8(const _Float16 * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16mf4_m))) +vfloat16mf4_t vluxei8(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m1))) +vfloat16m1_t vluxei16(const _Float16 * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m1_m))) +vfloat16m1_t vluxei16(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m2))) +vfloat16m2_t vluxei16(const _Float16 * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m2_m))) +vfloat16m2_t vluxei16(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m4))) +vfloat16m4_t vluxei16(const _Float16 * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m4_m))) +vfloat16m4_t vluxei16(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m8))) +vfloat16m8_t vluxei16(const _Float16 * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m8_m))) +vfloat16m8_t vluxei16(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16mf2))) +vfloat16mf2_t vluxei16(const _Float16 * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16mf2_m))) +vfloat16mf2_t vluxei16(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16mf4))) +vfloat16mf4_t vluxei16(const _Float16 * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16mf4_m))) +vfloat16mf4_t vluxei16(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16m1))) +vfloat16m1_t vluxei32(const _Float16 * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16m1_m))) +vfloat16m1_t vluxei32(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16m2))) +vfloat16m2_t vluxei32(const _Float16 * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16m2_m))) +vfloat16m2_t vluxei32(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16m4))) +vfloat16m4_t vluxei32(const _Float16 * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16m4_m))) +vfloat16m4_t vluxei32(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16mf2))) +vfloat16mf2_t vluxei32(const _Float16 * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16mf2_m))) +vfloat16mf2_t vluxei32(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16mf4))) +vfloat16mf4_t vluxei32(const _Float16 * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16mf4_m))) +vfloat16mf4_t vluxei32(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16m1))) +vfloat16m1_t vluxei64(const _Float16 * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16m1_m))) +vfloat16m1_t vluxei64(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16m2))) +vfloat16m2_t vluxei64(const _Float16 * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16m2_m))) +vfloat16m2_t vluxei64(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16mf2))) +vfloat16mf2_t vluxei64(const _Float16 * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16mf2_m))) +vfloat16mf2_t vluxei64(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16mf4))) +vfloat16mf4_t vluxei64(const _Float16 * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16mf4_m))) +vfloat16mf4_t vluxei64(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m1))) +vfloat16m1_t vloxei8(const _Float16 * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m1_m))) +vfloat16m1_t vloxei8(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m2))) +vfloat16m2_t vloxei8(const _Float16 * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m2_m))) +vfloat16m2_t vloxei8(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m4))) +vfloat16m4_t vloxei8(const _Float16 * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m4_m))) +vfloat16m4_t vloxei8(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m8))) +vfloat16m8_t vloxei8(const _Float16 * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m8_m))) +vfloat16m8_t vloxei8(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16mf2))) +vfloat16mf2_t vloxei8(const _Float16 * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16mf2_m))) +vfloat16mf2_t vloxei8(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16mf4))) +vfloat16mf4_t vloxei8(const _Float16 * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16mf4_m))) +vfloat16mf4_t vloxei8(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m1))) +vfloat16m1_t vloxei16(const _Float16 * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m1_m))) +vfloat16m1_t vloxei16(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m2))) +vfloat16m2_t vloxei16(const _Float16 * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m2_m))) +vfloat16m2_t vloxei16(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m4))) +vfloat16m4_t vloxei16(const _Float16 * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m4_m))) +vfloat16m4_t vloxei16(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m8))) +vfloat16m8_t vloxei16(const _Float16 * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m8_m))) +vfloat16m8_t vloxei16(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16mf2))) +vfloat16mf2_t vloxei16(const _Float16 * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16mf2_m))) +vfloat16mf2_t vloxei16(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16mf4))) +vfloat16mf4_t vloxei16(const _Float16 * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16mf4_m))) +vfloat16mf4_t vloxei16(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16m1))) +vfloat16m1_t vloxei32(const _Float16 * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16m1_m))) +vfloat16m1_t vloxei32(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16m2))) +vfloat16m2_t vloxei32(const _Float16 * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16m2_m))) +vfloat16m2_t vloxei32(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16m4))) +vfloat16m4_t vloxei32(const _Float16 * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16m4_m))) +vfloat16m4_t vloxei32(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16mf2))) +vfloat16mf2_t vloxei32(const _Float16 * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16mf2_m))) +vfloat16mf2_t vloxei32(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16mf4))) +vfloat16mf4_t vloxei32(const _Float16 * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16mf4_m))) +vfloat16mf4_t vloxei32(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16m1))) +vfloat16m1_t vloxei64(const _Float16 * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16m1_m))) +vfloat16m1_t vloxei64(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16m2))) +vfloat16m2_t vloxei64(const _Float16 * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16m2_m))) +vfloat16m2_t vloxei64(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16mf2))) +vfloat16mf2_t vloxei64(const _Float16 * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16mf2_m))) +vfloat16mf2_t vloxei64(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16mf4))) +vfloat16mf4_t vloxei64(const _Float16 * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16mf4_m))) +vfloat16mf4_t vloxei64(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f16m1))) +vfloat16m1_t vmv_v(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f16m2))) +vfloat16m2_t vmv_v(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f16m4))) +vfloat16m4_t vmv_v(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f16m8))) +vfloat16m8_t vmv_v(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f16mf2))) +vfloat16mf2_t vmv_v(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f16mf4))) +vfloat16mf4_t vmv_v(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m1))) +vfloat16m1_t vfadd(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m1_m))) +vfloat16m1_t vfadd(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m2))) +vfloat16m2_t vfadd(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m2_m))) +vfloat16m2_t vfadd(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m4))) +vfloat16m4_t vfadd(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m4_m))) +vfloat16m4_t vfadd(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m8))) +vfloat16m8_t vfadd(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m8_m))) +vfloat16m8_t vfadd(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16mf2))) +vfloat16mf2_t vfadd(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16mf2_m))) +vfloat16mf2_t vfadd(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16mf4))) +vfloat16mf4_t vfadd(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16mf4_m))) +vfloat16mf4_t vfadd(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m1))) +vfloat16m1_t vfadd(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m1_m))) +vfloat16m1_t vfadd(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m2))) +vfloat16m2_t vfadd(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m2_m))) +vfloat16m2_t vfadd(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m4))) +vfloat16m4_t vfadd(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m4_m))) +vfloat16m4_t vfadd(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m8))) +vfloat16m8_t vfadd(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m8_m))) +vfloat16m8_t vfadd(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16mf2))) +vfloat16mf2_t vfadd(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16mf2_m))) +vfloat16mf2_t vfadd(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16mf4))) +vfloat16mf4_t vfadd(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16mf4_m))) +vfloat16mf4_t vfadd(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m1))) +vfloat16m1_t vfsub(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m1_m))) +vfloat16m1_t vfsub(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m2))) +vfloat16m2_t vfsub(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m2_m))) +vfloat16m2_t vfsub(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m4))) +vfloat16m4_t vfsub(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m4_m))) +vfloat16m4_t vfsub(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m8))) +vfloat16m8_t vfsub(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m8_m))) +vfloat16m8_t vfsub(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16mf2))) +vfloat16mf2_t vfsub(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16mf2_m))) +vfloat16mf2_t vfsub(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16mf4))) +vfloat16mf4_t vfsub(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16mf4_m))) +vfloat16mf4_t vfsub(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m1))) +vfloat16m1_t vfsub(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m1_m))) +vfloat16m1_t vfsub(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m2))) +vfloat16m2_t vfsub(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m2_m))) +vfloat16m2_t vfsub(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m4))) +vfloat16m4_t vfsub(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m4_m))) +vfloat16m4_t vfsub(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m8))) +vfloat16m8_t vfsub(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m8_m))) +vfloat16m8_t vfsub(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16mf2))) +vfloat16mf2_t vfsub(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16mf2_m))) +vfloat16mf2_t vfsub(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16mf4))) +vfloat16mf4_t vfsub(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16mf4_m))) +vfloat16mf4_t vfsub(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m1))) +vfloat16m1_t vfrsub(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m1_m))) +vfloat16m1_t vfrsub(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m2))) +vfloat16m2_t vfrsub(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m2_m))) +vfloat16m2_t vfrsub(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m4))) +vfloat16m4_t vfrsub(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m4_m))) +vfloat16m4_t vfrsub(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m8))) +vfloat16m8_t vfrsub(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m8_m))) +vfloat16m8_t vfrsub(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16mf2))) +vfloat16mf2_t vfrsub(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16mf2_m))) +vfloat16mf2_t vfrsub(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16mf4))) +vfloat16mf4_t vfrsub(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16mf4_m))) +vfloat16mf4_t vfrsub(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m1))) +vfloat16m1_t vfmul(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m1_m))) +vfloat16m1_t vfmul(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m2))) +vfloat16m2_t vfmul(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m2_m))) +vfloat16m2_t vfmul(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m4))) +vfloat16m4_t vfmul(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m4_m))) +vfloat16m4_t vfmul(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m8))) +vfloat16m8_t vfmul(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m8_m))) +vfloat16m8_t vfmul(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16mf2))) +vfloat16mf2_t vfmul(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16mf2_m))) +vfloat16mf2_t vfmul(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16mf4))) +vfloat16mf4_t vfmul(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16mf4_m))) +vfloat16mf4_t vfmul(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m1))) +vfloat16m1_t vfmul(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m1_m))) +vfloat16m1_t vfmul(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m2))) +vfloat16m2_t vfmul(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m2_m))) +vfloat16m2_t vfmul(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m4))) +vfloat16m4_t vfmul(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m4_m))) +vfloat16m4_t vfmul(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m8))) +vfloat16m8_t vfmul(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m8_m))) +vfloat16m8_t vfmul(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16mf2))) +vfloat16mf2_t vfmul(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16mf2_m))) +vfloat16mf2_t vfmul(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16mf4))) +vfloat16mf4_t vfmul(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16mf4_m))) +vfloat16mf4_t vfmul(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m1))) +vfloat16m1_t vfdiv(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m1_m))) +vfloat16m1_t vfdiv(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m2))) +vfloat16m2_t vfdiv(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m2_m))) +vfloat16m2_t vfdiv(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m4))) +vfloat16m4_t vfdiv(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m4_m))) +vfloat16m4_t vfdiv(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m8))) +vfloat16m8_t vfdiv(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m8_m))) +vfloat16m8_t vfdiv(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16mf2))) +vfloat16mf2_t vfdiv(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16mf2_m))) +vfloat16mf2_t vfdiv(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16mf4))) +vfloat16mf4_t vfdiv(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16mf4_m))) +vfloat16mf4_t vfdiv(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m1))) +vfloat16m1_t vfdiv(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m1_m))) +vfloat16m1_t vfdiv(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m2))) +vfloat16m2_t vfdiv(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m2_m))) +vfloat16m2_t vfdiv(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m4))) +vfloat16m4_t vfdiv(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m4_m))) +vfloat16m4_t vfdiv(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m8))) +vfloat16m8_t vfdiv(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m8_m))) +vfloat16m8_t vfdiv(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16mf2))) +vfloat16mf2_t vfdiv(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16mf2_m))) +vfloat16mf2_t vfdiv(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16mf4))) +vfloat16mf4_t vfdiv(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16mf4_m))) +vfloat16mf4_t vfdiv(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m1))) +vfloat16m1_t vfrdiv(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m1_m))) +vfloat16m1_t vfrdiv(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m2))) +vfloat16m2_t vfrdiv(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m2_m))) +vfloat16m2_t vfrdiv(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m4))) +vfloat16m4_t vfrdiv(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m4_m))) +vfloat16m4_t vfrdiv(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m8))) +vfloat16m8_t vfrdiv(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m8_m))) +vfloat16m8_t vfrdiv(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16mf2))) +vfloat16mf2_t vfrdiv(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16mf2_m))) +vfloat16mf2_t vfrdiv(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16mf4))) +vfloat16mf4_t vfrdiv(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16mf4_m))) +vfloat16mf4_t vfrdiv(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m1))) +vfloat16m1_t vfmacc(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m1_m))) +vfloat16m1_t vfmacc(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m2))) +vfloat16m2_t vfmacc(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m2_m))) +vfloat16m2_t vfmacc(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m4))) +vfloat16m4_t vfmacc(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m4_m))) +vfloat16m4_t vfmacc(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m8))) +vfloat16m8_t vfmacc(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m8_m))) +vfloat16m8_t vfmacc(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16mf2))) +vfloat16mf2_t vfmacc(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16mf2_m))) +vfloat16mf2_t vfmacc(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16mf4))) +vfloat16mf4_t vfmacc(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16mf4_m))) +vfloat16mf4_t vfmacc(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m1))) +vfloat16m1_t vfmacc(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m1_m))) +vfloat16m1_t vfmacc(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m2))) +vfloat16m2_t vfmacc(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m2_m))) +vfloat16m2_t vfmacc(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m4))) +vfloat16m4_t vfmacc(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m4_m))) +vfloat16m4_t vfmacc(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m8))) +vfloat16m8_t vfmacc(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m8_m))) +vfloat16m8_t vfmacc(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16mf2))) +vfloat16mf2_t vfmacc(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16mf2_m))) +vfloat16mf2_t vfmacc(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16mf4))) +vfloat16mf4_t vfmacc(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16mf4_m))) +vfloat16mf4_t vfmacc(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m1))) +vfloat16m1_t vfnmacc(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m1_m))) +vfloat16m1_t vfnmacc(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m2))) +vfloat16m2_t vfnmacc(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m2_m))) +vfloat16m2_t vfnmacc(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m4))) +vfloat16m4_t vfnmacc(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m4_m))) +vfloat16m4_t vfnmacc(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m8))) +vfloat16m8_t vfnmacc(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m8_m))) +vfloat16m8_t vfnmacc(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16mf2))) +vfloat16mf2_t vfnmacc(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16mf2_m))) +vfloat16mf2_t vfnmacc(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16mf4))) +vfloat16mf4_t vfnmacc(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16mf4_m))) +vfloat16mf4_t vfnmacc(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m1))) +vfloat16m1_t vfnmacc(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m1_m))) +vfloat16m1_t vfnmacc(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m2))) +vfloat16m2_t vfnmacc(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m2_m))) +vfloat16m2_t vfnmacc(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m4))) +vfloat16m4_t vfnmacc(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m4_m))) +vfloat16m4_t vfnmacc(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m8))) +vfloat16m8_t vfnmacc(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m8_m))) +vfloat16m8_t vfnmacc(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16mf2))) +vfloat16mf2_t vfnmacc(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16mf2_m))) +vfloat16mf2_t vfnmacc(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16mf4))) +vfloat16mf4_t vfnmacc(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16mf4_m))) +vfloat16mf4_t vfnmacc(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m1))) +vfloat16m1_t vfmsac(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m1_m))) +vfloat16m1_t vfmsac(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m2))) +vfloat16m2_t vfmsac(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m2_m))) +vfloat16m2_t vfmsac(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m4))) +vfloat16m4_t vfmsac(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m4_m))) +vfloat16m4_t vfmsac(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m8))) +vfloat16m8_t vfmsac(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m8_m))) +vfloat16m8_t vfmsac(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16mf2))) +vfloat16mf2_t vfmsac(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16mf2_m))) +vfloat16mf2_t vfmsac(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16mf4))) +vfloat16mf4_t vfmsac(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16mf4_m))) +vfloat16mf4_t vfmsac(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m1))) +vfloat16m1_t vfmsac(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m1_m))) +vfloat16m1_t vfmsac(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m2))) +vfloat16m2_t vfmsac(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m2_m))) +vfloat16m2_t vfmsac(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m4))) +vfloat16m4_t vfmsac(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m4_m))) +vfloat16m4_t vfmsac(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m8))) +vfloat16m8_t vfmsac(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m8_m))) +vfloat16m8_t vfmsac(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16mf2))) +vfloat16mf2_t vfmsac(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16mf2_m))) +vfloat16mf2_t vfmsac(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16mf4))) +vfloat16mf4_t vfmsac(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16mf4_m))) +vfloat16mf4_t vfmsac(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m1))) +vfloat16m1_t vfnmsac(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m1_m))) +vfloat16m1_t vfnmsac(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m2))) +vfloat16m2_t vfnmsac(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m2_m))) +vfloat16m2_t vfnmsac(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m4))) +vfloat16m4_t vfnmsac(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m4_m))) +vfloat16m4_t vfnmsac(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m8))) +vfloat16m8_t vfnmsac(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m8_m))) +vfloat16m8_t vfnmsac(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16mf2))) +vfloat16mf2_t vfnmsac(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16mf2_m))) +vfloat16mf2_t vfnmsac(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16mf4))) +vfloat16mf4_t vfnmsac(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16mf4_m))) +vfloat16mf4_t vfnmsac(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m1))) +vfloat16m1_t vfnmsac(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m1_m))) +vfloat16m1_t vfnmsac(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m2))) +vfloat16m2_t vfnmsac(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m2_m))) +vfloat16m2_t vfnmsac(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m4))) +vfloat16m4_t vfnmsac(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m4_m))) +vfloat16m4_t vfnmsac(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m8))) +vfloat16m8_t vfnmsac(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m8_m))) +vfloat16m8_t vfnmsac(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16mf2))) +vfloat16mf2_t vfnmsac(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16mf2_m))) +vfloat16mf2_t vfnmsac(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16mf4))) +vfloat16mf4_t vfnmsac(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16mf4_m))) +vfloat16mf4_t vfnmsac(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m1))) +vfloat16m1_t vfmadd(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m1_m))) +vfloat16m1_t vfmadd(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m2))) +vfloat16m2_t vfmadd(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m2_m))) +vfloat16m2_t vfmadd(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m4))) +vfloat16m4_t vfmadd(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m4_m))) +vfloat16m4_t vfmadd(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m8))) +vfloat16m8_t vfmadd(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m8_m))) +vfloat16m8_t vfmadd(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16mf2))) +vfloat16mf2_t vfmadd(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16mf2_m))) +vfloat16mf2_t vfmadd(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16mf4))) +vfloat16mf4_t vfmadd(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16mf4_m))) +vfloat16mf4_t vfmadd(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m1))) +vfloat16m1_t vfmadd(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m1_m))) +vfloat16m1_t vfmadd(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m2))) +vfloat16m2_t vfmadd(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m2_m))) +vfloat16m2_t vfmadd(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m4))) +vfloat16m4_t vfmadd(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m4_m))) +vfloat16m4_t vfmadd(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m8))) +vfloat16m8_t vfmadd(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m8_m))) +vfloat16m8_t vfmadd(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16mf2))) +vfloat16mf2_t vfmadd(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16mf2_m))) +vfloat16mf2_t vfmadd(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16mf4))) +vfloat16mf4_t vfmadd(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16mf4_m))) +vfloat16mf4_t vfmadd(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m1))) +vfloat16m1_t vfnmadd(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m1_m))) +vfloat16m1_t vfnmadd(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m2))) +vfloat16m2_t vfnmadd(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m2_m))) +vfloat16m2_t vfnmadd(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m4))) +vfloat16m4_t vfnmadd(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m4_m))) +vfloat16m4_t vfnmadd(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m8))) +vfloat16m8_t vfnmadd(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m8_m))) +vfloat16m8_t vfnmadd(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16mf2))) +vfloat16mf2_t vfnmadd(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16mf2_m))) +vfloat16mf2_t vfnmadd(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16mf4))) +vfloat16mf4_t vfnmadd(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16mf4_m))) +vfloat16mf4_t vfnmadd(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m1))) +vfloat16m1_t vfnmadd(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m1_m))) +vfloat16m1_t vfnmadd(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m2))) +vfloat16m2_t vfnmadd(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m2_m))) +vfloat16m2_t vfnmadd(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m4))) +vfloat16m4_t vfnmadd(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m4_m))) +vfloat16m4_t vfnmadd(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m8))) +vfloat16m8_t vfnmadd(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m8_m))) +vfloat16m8_t vfnmadd(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16mf2))) +vfloat16mf2_t vfnmadd(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16mf2_m))) +vfloat16mf2_t vfnmadd(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16mf4))) +vfloat16mf4_t vfnmadd(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16mf4_m))) +vfloat16mf4_t vfnmadd(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m1))) +vfloat16m1_t vfmsub(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m1_m))) +vfloat16m1_t vfmsub(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m2))) +vfloat16m2_t vfmsub(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m2_m))) +vfloat16m2_t vfmsub(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m4))) +vfloat16m4_t vfmsub(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m4_m))) +vfloat16m4_t vfmsub(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m8))) +vfloat16m8_t vfmsub(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m8_m))) +vfloat16m8_t vfmsub(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16mf2))) +vfloat16mf2_t vfmsub(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16mf2_m))) +vfloat16mf2_t vfmsub(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16mf4))) +vfloat16mf4_t vfmsub(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16mf4_m))) +vfloat16mf4_t vfmsub(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m1))) +vfloat16m1_t vfmsub(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m1_m))) +vfloat16m1_t vfmsub(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m2))) +vfloat16m2_t vfmsub(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m2_m))) +vfloat16m2_t vfmsub(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m4))) +vfloat16m4_t vfmsub(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m4_m))) +vfloat16m4_t vfmsub(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m8))) +vfloat16m8_t vfmsub(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m8_m))) +vfloat16m8_t vfmsub(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16mf2))) +vfloat16mf2_t vfmsub(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16mf2_m))) +vfloat16mf2_t vfmsub(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16mf4))) +vfloat16mf4_t vfmsub(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16mf4_m))) +vfloat16mf4_t vfmsub(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m1))) +vfloat16m1_t vfnmsub(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m1_m))) +vfloat16m1_t vfnmsub(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m2))) +vfloat16m2_t vfnmsub(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m2_m))) +vfloat16m2_t vfnmsub(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m4))) +vfloat16m4_t vfnmsub(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m4_m))) +vfloat16m4_t vfnmsub(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m8))) +vfloat16m8_t vfnmsub(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m8_m))) +vfloat16m8_t vfnmsub(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16mf2))) +vfloat16mf2_t vfnmsub(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16mf2_m))) +vfloat16mf2_t vfnmsub(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16mf4))) +vfloat16mf4_t vfnmsub(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16mf4_m))) +vfloat16mf4_t vfnmsub(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m1))) +vfloat16m1_t vfnmsub(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m1_m))) +vfloat16m1_t vfnmsub(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m2))) +vfloat16m2_t vfnmsub(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m2_m))) +vfloat16m2_t vfnmsub(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m4))) +vfloat16m4_t vfnmsub(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m4_m))) +vfloat16m4_t vfnmsub(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m8))) +vfloat16m8_t vfnmsub(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m8_m))) +vfloat16m8_t vfnmsub(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16mf2))) +vfloat16mf2_t vfnmsub(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16mf2_m))) +vfloat16mf2_t vfnmsub(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16mf4))) +vfloat16mf4_t vfnmsub(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16mf4_m))) +vfloat16mf4_t vfnmsub(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m1))) +vfloat16m1_t vfmin(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m1_m))) +vfloat16m1_t vfmin(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m2))) +vfloat16m2_t vfmin(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m2_m))) +vfloat16m2_t vfmin(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m4))) +vfloat16m4_t vfmin(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m4_m))) +vfloat16m4_t vfmin(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m8))) +vfloat16m8_t vfmin(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m8_m))) +vfloat16m8_t vfmin(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16mf2))) +vfloat16mf2_t vfmin(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16mf2_m))) +vfloat16mf2_t vfmin(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16mf4))) +vfloat16mf4_t vfmin(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16mf4_m))) +vfloat16mf4_t vfmin(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m1))) +vfloat16m1_t vfmin(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m1_m))) +vfloat16m1_t vfmin(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m2))) +vfloat16m2_t vfmin(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m2_m))) +vfloat16m2_t vfmin(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m4))) +vfloat16m4_t vfmin(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m4_m))) +vfloat16m4_t vfmin(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m8))) +vfloat16m8_t vfmin(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m8_m))) +vfloat16m8_t vfmin(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16mf2))) +vfloat16mf2_t vfmin(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16mf2_m))) +vfloat16mf2_t vfmin(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16mf4))) +vfloat16mf4_t vfmin(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16mf4_m))) +vfloat16mf4_t vfmin(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m1))) +vfloat16m1_t vfmax(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m1_m))) +vfloat16m1_t vfmax(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m2))) +vfloat16m2_t vfmax(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m2_m))) +vfloat16m2_t vfmax(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m4))) +vfloat16m4_t vfmax(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m4_m))) +vfloat16m4_t vfmax(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m8))) +vfloat16m8_t vfmax(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m8_m))) +vfloat16m8_t vfmax(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16mf2))) +vfloat16mf2_t vfmax(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16mf2_m))) +vfloat16mf2_t vfmax(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16mf4))) +vfloat16mf4_t vfmax(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16mf4_m))) +vfloat16mf4_t vfmax(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m1))) +vfloat16m1_t vfmax(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m1_m))) +vfloat16m1_t vfmax(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m2))) +vfloat16m2_t vfmax(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m2_m))) +vfloat16m2_t vfmax(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m4))) +vfloat16m4_t vfmax(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m4_m))) +vfloat16m4_t vfmax(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m8))) +vfloat16m8_t vfmax(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m8_m))) +vfloat16m8_t vfmax(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16mf2))) +vfloat16mf2_t vfmax(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16mf2_m))) +vfloat16mf2_t vfmax(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16mf4))) +vfloat16mf4_t vfmax(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16mf4_m))) +vfloat16mf4_t vfmax(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m1))) +vfloat16m1_t vfsgnj(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m1_m))) +vfloat16m1_t vfsgnj(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m2))) +vfloat16m2_t vfsgnj(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m2_m))) +vfloat16m2_t vfsgnj(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m4))) +vfloat16m4_t vfsgnj(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m4_m))) +vfloat16m4_t vfsgnj(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m8))) +vfloat16m8_t vfsgnj(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m8_m))) +vfloat16m8_t vfsgnj(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16mf2))) +vfloat16mf2_t vfsgnj(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16mf2_m))) +vfloat16mf2_t vfsgnj(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16mf4))) +vfloat16mf4_t vfsgnj(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16mf4_m))) +vfloat16mf4_t vfsgnj(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m1))) +vfloat16m1_t vfsgnj(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m1_m))) +vfloat16m1_t vfsgnj(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m2))) +vfloat16m2_t vfsgnj(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m2_m))) +vfloat16m2_t vfsgnj(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m4))) +vfloat16m4_t vfsgnj(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m4_m))) +vfloat16m4_t vfsgnj(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m8))) +vfloat16m8_t vfsgnj(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m8_m))) +vfloat16m8_t vfsgnj(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16mf2))) +vfloat16mf2_t vfsgnj(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16mf2_m))) +vfloat16mf2_t vfsgnj(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16mf4))) +vfloat16mf4_t vfsgnj(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16mf4_m))) +vfloat16mf4_t vfsgnj(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m1))) +vfloat16m1_t vfsgnjn(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m1_m))) +vfloat16m1_t vfsgnjn(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m2))) +vfloat16m2_t vfsgnjn(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m2_m))) +vfloat16m2_t vfsgnjn(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m4))) +vfloat16m4_t vfsgnjn(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m4_m))) +vfloat16m4_t vfsgnjn(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m8))) +vfloat16m8_t vfsgnjn(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m8_m))) +vfloat16m8_t vfsgnjn(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16mf2))) +vfloat16mf2_t vfsgnjn(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16mf2_m))) +vfloat16mf2_t vfsgnjn(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16mf4))) +vfloat16mf4_t vfsgnjn(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16mf4_m))) +vfloat16mf4_t vfsgnjn(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m1))) +vfloat16m1_t vfsgnjn(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m1_m))) +vfloat16m1_t vfsgnjn(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m2))) +vfloat16m2_t vfsgnjn(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m2_m))) +vfloat16m2_t vfsgnjn(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m4))) +vfloat16m4_t vfsgnjn(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m4_m))) +vfloat16m4_t vfsgnjn(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m8))) +vfloat16m8_t vfsgnjn(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m8_m))) +vfloat16m8_t vfsgnjn(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16mf2))) +vfloat16mf2_t vfsgnjn(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16mf2_m))) +vfloat16mf2_t vfsgnjn(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16mf4))) +vfloat16mf4_t vfsgnjn(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16mf4_m))) +vfloat16mf4_t vfsgnjn(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m1))) +vfloat16m1_t vfsgnjx(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m1_m))) +vfloat16m1_t vfsgnjx(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m2))) +vfloat16m2_t vfsgnjx(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m2_m))) +vfloat16m2_t vfsgnjx(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m4))) +vfloat16m4_t vfsgnjx(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m4_m))) +vfloat16m4_t vfsgnjx(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m8))) +vfloat16m8_t vfsgnjx(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m8_m))) +vfloat16m8_t vfsgnjx(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16mf2))) +vfloat16mf2_t vfsgnjx(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16mf2_m))) +vfloat16mf2_t vfsgnjx(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16mf4))) +vfloat16mf4_t vfsgnjx(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16mf4_m))) +vfloat16mf4_t vfsgnjx(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m1))) +vfloat16m1_t vfsgnjx(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m1_m))) +vfloat16m1_t vfsgnjx(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m2))) +vfloat16m2_t vfsgnjx(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m2_m))) +vfloat16m2_t vfsgnjx(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m4))) +vfloat16m4_t vfsgnjx(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m4_m))) +vfloat16m4_t vfsgnjx(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m8))) +vfloat16m8_t vfsgnjx(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m8_m))) +vfloat16m8_t vfsgnjx(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16mf2))) +vfloat16mf2_t vfsgnjx(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16mf2_m))) +vfloat16mf2_t vfsgnjx(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16mf4))) +vfloat16mf4_t vfsgnjx(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16mf4_m))) +vfloat16mf4_t vfsgnjx(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m1))) +vfloat16m1_t vfabs(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m1_m))) +vfloat16m1_t vfabs(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m2))) +vfloat16m2_t vfabs(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m2_m))) +vfloat16m2_t vfabs(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m4))) +vfloat16m4_t vfabs(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m4_m))) +vfloat16m4_t vfabs(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m8))) +vfloat16m8_t vfabs(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m8_m))) +vfloat16m8_t vfabs(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16mf2))) +vfloat16mf2_t vfabs(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16mf2_m))) +vfloat16mf2_t vfabs(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16mf4))) +vfloat16mf4_t vfabs(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16mf4_m))) +vfloat16mf4_t vfabs(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m1_b16))) +vbool16_t vmfeq(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m1_b16_m))) +vbool16_t vmfeq(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m2_b8))) +vbool8_t vmfeq(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m2_b8_m))) +vbool8_t vmfeq(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m4_b4))) +vbool4_t vmfeq(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m4_b4_m))) +vbool4_t vmfeq(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m8_b2))) +vbool2_t vmfeq(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m8_b2_m))) +vbool2_t vmfeq(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16mf2_b32))) +vbool32_t vmfeq(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16mf2_b32_m))) +vbool32_t vmfeq(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16mf4_b64))) +vbool64_t vmfeq(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16mf4_b64_m))) +vbool64_t vmfeq(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m1_b16))) +vbool16_t vmfeq(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m1_b16_m))) +vbool16_t vmfeq(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m2_b8))) +vbool8_t vmfeq(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m2_b8_m))) +vbool8_t vmfeq(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m4_b4))) +vbool4_t vmfeq(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m4_b4_m))) +vbool4_t vmfeq(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m8_b2))) +vbool2_t vmfeq(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m8_b2_m))) +vbool2_t vmfeq(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16mf2_b32))) +vbool32_t vmfeq(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16mf2_b32_m))) +vbool32_t vmfeq(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16mf4_b64))) +vbool64_t vmfeq(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16mf4_b64_m))) +vbool64_t vmfeq(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m1_b16))) +vbool16_t vmfne(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m1_b16_m))) +vbool16_t vmfne(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m2_b8))) +vbool8_t vmfne(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m2_b8_m))) +vbool8_t vmfne(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m4_b4))) +vbool4_t vmfne(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m4_b4_m))) +vbool4_t vmfne(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m8_b2))) +vbool2_t vmfne(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m8_b2_m))) +vbool2_t vmfne(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16mf2_b32))) +vbool32_t vmfne(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16mf2_b32_m))) +vbool32_t vmfne(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16mf4_b64))) +vbool64_t vmfne(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16mf4_b64_m))) +vbool64_t vmfne(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m1_b16))) +vbool16_t vmfne(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m1_b16_m))) +vbool16_t vmfne(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m2_b8))) +vbool8_t vmfne(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m2_b8_m))) +vbool8_t vmfne(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m4_b4))) +vbool4_t vmfne(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m4_b4_m))) +vbool4_t vmfne(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m8_b2))) +vbool2_t vmfne(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m8_b2_m))) +vbool2_t vmfne(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16mf2_b32))) +vbool32_t vmfne(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16mf2_b32_m))) +vbool32_t vmfne(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16mf4_b64))) +vbool64_t vmfne(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16mf4_b64_m))) +vbool64_t vmfne(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m1_b16))) +vbool16_t vmflt(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m1_b16_m))) +vbool16_t vmflt(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m2_b8))) +vbool8_t vmflt(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m2_b8_m))) +vbool8_t vmflt(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m4_b4))) +vbool4_t vmflt(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m4_b4_m))) +vbool4_t vmflt(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m8_b2))) +vbool2_t vmflt(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m8_b2_m))) +vbool2_t vmflt(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16mf2_b32))) +vbool32_t vmflt(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16mf2_b32_m))) +vbool32_t vmflt(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16mf4_b64))) +vbool64_t vmflt(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16mf4_b64_m))) +vbool64_t vmflt(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m1_b16))) +vbool16_t vmflt(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m1_b16_m))) +vbool16_t vmflt(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m2_b8))) +vbool8_t vmflt(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m2_b8_m))) +vbool8_t vmflt(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m4_b4))) +vbool4_t vmflt(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m4_b4_m))) +vbool4_t vmflt(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m8_b2))) +vbool2_t vmflt(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m8_b2_m))) +vbool2_t vmflt(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16mf2_b32))) +vbool32_t vmflt(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16mf2_b32_m))) +vbool32_t vmflt(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16mf4_b64))) +vbool64_t vmflt(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16mf4_b64_m))) +vbool64_t vmflt(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m1_b16))) +vbool16_t vmfle(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m1_b16_m))) +vbool16_t vmfle(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m2_b8))) +vbool8_t vmfle(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m2_b8_m))) +vbool8_t vmfle(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m4_b4))) +vbool4_t vmfle(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m4_b4_m))) +vbool4_t vmfle(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m8_b2))) +vbool2_t vmfle(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m8_b2_m))) +vbool2_t vmfle(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16mf2_b32))) +vbool32_t vmfle(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16mf2_b32_m))) +vbool32_t vmfle(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16mf4_b64))) +vbool64_t vmfle(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16mf4_b64_m))) +vbool64_t vmfle(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m1_b16))) +vbool16_t vmfle(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m1_b16_m))) +vbool16_t vmfle(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m2_b8))) +vbool8_t vmfle(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m2_b8_m))) +vbool8_t vmfle(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m4_b4))) +vbool4_t vmfle(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m4_b4_m))) +vbool4_t vmfle(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m8_b2))) +vbool2_t vmfle(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m8_b2_m))) +vbool2_t vmfle(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16mf2_b32))) +vbool32_t vmfle(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16mf2_b32_m))) +vbool32_t vmfle(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16mf4_b64))) +vbool64_t vmfle(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16mf4_b64_m))) +vbool64_t vmfle(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m1_b16))) +vbool16_t vmfgt(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m1_b16_m))) +vbool16_t vmfgt(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m2_b8))) +vbool8_t vmfgt(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m2_b8_m))) +vbool8_t vmfgt(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m4_b4))) +vbool4_t vmfgt(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m4_b4_m))) +vbool4_t vmfgt(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m8_b2))) +vbool2_t vmfgt(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m8_b2_m))) +vbool2_t vmfgt(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16mf2_b32))) +vbool32_t vmfgt(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16mf2_b32_m))) +vbool32_t vmfgt(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16mf4_b64))) +vbool64_t vmfgt(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16mf4_b64_m))) +vbool64_t vmfgt(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m1_b16))) +vbool16_t vmfgt(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m1_b16_m))) +vbool16_t vmfgt(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m2_b8))) +vbool8_t vmfgt(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m2_b8_m))) +vbool8_t vmfgt(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m4_b4))) +vbool4_t vmfgt(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m4_b4_m))) +vbool4_t vmfgt(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m8_b2))) +vbool2_t vmfgt(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m8_b2_m))) +vbool2_t vmfgt(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16mf2_b32))) +vbool32_t vmfgt(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16mf2_b32_m))) +vbool32_t vmfgt(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16mf4_b64))) +vbool64_t vmfgt(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16mf4_b64_m))) +vbool64_t vmfgt(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m1_b16))) +vbool16_t vmfge(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m1_b16_m))) +vbool16_t vmfge(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m2_b8))) +vbool8_t vmfge(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m2_b8_m))) +vbool8_t vmfge(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m4_b4))) +vbool4_t vmfge(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m4_b4_m))) +vbool4_t vmfge(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m8_b2))) +vbool2_t vmfge(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m8_b2_m))) +vbool2_t vmfge(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16mf2_b32))) +vbool32_t vmfge(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16mf2_b32_m))) +vbool32_t vmfge(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16mf4_b64))) +vbool64_t vmfge(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16mf4_b64_m))) +vbool64_t vmfge(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m1_b16))) +vbool16_t vmfge(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m1_b16_m))) +vbool16_t vmfge(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m2_b8))) +vbool8_t vmfge(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m2_b8_m))) +vbool8_t vmfge(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m4_b4))) +vbool4_t vmfge(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m4_b4_m))) +vbool4_t vmfge(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m8_b2))) +vbool2_t vmfge(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m8_b2_m))) +vbool2_t vmfge(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16mf2_b32))) +vbool32_t vmfge(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16mf2_b32_m))) +vbool32_t vmfge(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16mf4_b64))) +vbool64_t vmfge(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16mf4_b64_m))) +vbool64_t vmfge(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f16m1))) +vfloat16m1_t vmerge(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f16m2))) +vfloat16m2_t vmerge(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f16m4))) +vfloat16m4_t vmerge(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f16m8))) +vfloat16m8_t vmerge(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f16mf2))) +vfloat16mf2_t vmerge(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f16mf4))) +vfloat16mf4_t vmerge(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f16m1))) +vfloat16m1_t vfmerge(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f16m2))) +vfloat16m2_t vfmerge(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f16m4))) +vfloat16m4_t vfmerge(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f16m8))) +vfloat16m8_t vfmerge(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f16mf2))) +vfloat16mf2_t vfmerge(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f16mf4))) +vfloat16mf4_t vfmerge(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m1_f16m1))) +vfloat16m1_t vfredmax(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m1_f16m1_m))) +vfloat16m1_t vfredmax(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m2_f16m1))) +vfloat16m1_t vfredmax(vfloat16m1_t op0, vfloat16m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m2_f16m1_m))) +vfloat16m1_t vfredmax(vbool8_t op0, vfloat16m1_t op1, vfloat16m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m4_f16m1))) +vfloat16m1_t vfredmax(vfloat16m1_t op0, vfloat16m4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m4_f16m1_m))) +vfloat16m1_t vfredmax(vbool4_t op0, vfloat16m1_t op1, vfloat16m4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m8_f16m1))) +vfloat16m1_t vfredmax(vfloat16m1_t op0, vfloat16m8_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m8_f16m1_m))) +vfloat16m1_t vfredmax(vbool2_t op0, vfloat16m1_t op1, vfloat16m8_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16mf2_f16m1))) +vfloat16m1_t vfredmax(vfloat16m1_t op0, vfloat16mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16mf2_f16m1_m))) +vfloat16m1_t vfredmax(vbool32_t op0, vfloat16m1_t op1, vfloat16mf2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16mf4_f16m1))) +vfloat16m1_t vfredmax(vfloat16m1_t op0, vfloat16mf4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16mf4_f16m1_m))) +vfloat16m1_t vfredmax(vbool64_t op0, vfloat16m1_t op1, vfloat16mf4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m1_f16m1))) +vfloat16m1_t vfredmin(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m1_f16m1_m))) +vfloat16m1_t vfredmin(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m2_f16m1))) +vfloat16m1_t vfredmin(vfloat16m1_t op0, vfloat16m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m2_f16m1_m))) +vfloat16m1_t vfredmin(vbool8_t op0, vfloat16m1_t op1, vfloat16m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m4_f16m1))) +vfloat16m1_t vfredmin(vfloat16m1_t op0, vfloat16m4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m4_f16m1_m))) +vfloat16m1_t vfredmin(vbool4_t op0, vfloat16m1_t op1, vfloat16m4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m8_f16m1))) +vfloat16m1_t vfredmin(vfloat16m1_t op0, vfloat16m8_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m8_f16m1_m))) +vfloat16m1_t vfredmin(vbool2_t op0, vfloat16m1_t op1, vfloat16m8_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16mf2_f16m1))) +vfloat16m1_t vfredmin(vfloat16m1_t op0, vfloat16mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16mf2_f16m1_m))) +vfloat16m1_t vfredmin(vbool32_t op0, vfloat16m1_t op1, vfloat16mf2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16mf4_f16m1))) +vfloat16m1_t vfredmin(vfloat16m1_t op0, vfloat16mf4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16mf4_f16m1_m))) +vfloat16m1_t vfredmin(vbool64_t op0, vfloat16m1_t op1, vfloat16mf4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m1_f16m1))) +vfloat16m1_t vfredsum(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m1_f16m1_m))) +vfloat16m1_t vfredsum(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m2_f16m1))) +vfloat16m1_t vfredsum(vfloat16m1_t op0, vfloat16m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m2_f16m1_m))) +vfloat16m1_t vfredsum(vbool8_t op0, vfloat16m1_t op1, vfloat16m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m4_f16m1))) +vfloat16m1_t vfredsum(vfloat16m1_t op0, vfloat16m4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m4_f16m1_m))) +vfloat16m1_t vfredsum(vbool4_t op0, vfloat16m1_t op1, vfloat16m4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m8_f16m1))) +vfloat16m1_t vfredsum(vfloat16m1_t op0, vfloat16m8_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m8_f16m1_m))) +vfloat16m1_t vfredsum(vbool2_t op0, vfloat16m1_t op1, vfloat16m8_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16mf2_f16m1))) +vfloat16m1_t vfredsum(vfloat16m1_t op0, vfloat16mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16mf2_f16m1_m))) +vfloat16m1_t vfredsum(vbool32_t op0, vfloat16m1_t op1, vfloat16mf2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16mf4_f16m1))) +vfloat16m1_t vfredsum(vfloat16m1_t op0, vfloat16mf4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16mf4_f16m1_m))) +vfloat16m1_t vfredsum(vbool64_t op0, vfloat16m1_t op1, vfloat16mf4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m1_f16m1))) +vfloat16m1_t vfredosum(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m1_f16m1_m))) +vfloat16m1_t vfredosum(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m2_f16m1))) +vfloat16m1_t vfredosum(vfloat16m1_t op0, vfloat16m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m2_f16m1_m))) +vfloat16m1_t vfredosum(vbool8_t op0, vfloat16m1_t op1, vfloat16m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m4_f16m1))) +vfloat16m1_t vfredosum(vfloat16m1_t op0, vfloat16m4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m4_f16m1_m))) +vfloat16m1_t vfredosum(vbool4_t op0, vfloat16m1_t op1, vfloat16m4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m8_f16m1))) +vfloat16m1_t vfredosum(vfloat16m1_t op0, vfloat16m8_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m8_f16m1_m))) +vfloat16m1_t vfredosum(vbool2_t op0, vfloat16m1_t op1, vfloat16m8_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16mf2_f16m1))) +vfloat16m1_t vfredosum(vfloat16m1_t op0, vfloat16mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16mf2_f16m1_m))) +vfloat16m1_t vfredosum(vbool32_t op0, vfloat16m1_t op1, vfloat16mf2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16mf4_f16m1))) +vfloat16m1_t vfredosum(vfloat16m1_t op0, vfloat16mf4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16mf4_f16m1_m))) +vfloat16m1_t vfredosum(vbool64_t op0, vfloat16m1_t op1, vfloat16mf4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m1))) +void vsuxei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m1_m))) +void vsuxei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m2))) +void vsuxei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m2_m))) +void vsuxei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m4))) +void vsuxei8(_Float16 * op0, vuint8m2_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m4_m))) +void vsuxei8(vbool4_t op0, _Float16 * op1, vuint8m2_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m8))) +void vsuxei8(_Float16 * op0, vuint8m4_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m8_m))) +void vsuxei8(vbool2_t op0, _Float16 * op1, vuint8m4_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16mf2))) +void vsuxei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16mf2_m))) +void vsuxei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16mf4))) +void vsuxei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16mf4_m))) +void vsuxei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m1))) +void vsuxei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m1_m))) +void vsuxei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m2))) +void vsuxei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m2_m))) +void vsuxei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m4))) +void vsuxei16(_Float16 * op0, vuint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m4_m))) +void vsuxei16(vbool4_t op0, _Float16 * op1, vuint16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m8))) +void vsuxei16(_Float16 * op0, vuint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m8_m))) +void vsuxei16(vbool2_t op0, _Float16 * op1, vuint16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16mf2))) +void vsuxei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16mf2_m))) +void vsuxei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16mf4))) +void vsuxei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16mf4_m))) +void vsuxei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f16m1_f16))) +_Float16 vfmv_f(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f16m2_f16))) +_Float16 vfmv_f(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f16m4_f16))) +_Float16 vfmv_f(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f16m8_f16))) +_Float16 vfmv_f(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f16mf2_f16))) +_Float16 vfmv_f(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f16mf4_f16))) +_Float16 vfmv_f(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f16m1))) +vfloat16m1_t vfmv_s(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f16m2))) +vfloat16m2_t vfmv_s(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f16m4))) +vfloat16m4_t vfmv_s(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f16m8))) +vfloat16m8_t vfmv_s(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f16mf2))) +vfloat16mf2_t vfmv_s(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f16mf4))) +vfloat16mf4_t vfmv_s(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m1))) +vfloat16m1_t vslideup(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m1_m))) +vfloat16m1_t vslideup(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m2))) +vfloat16m2_t vslideup(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m2_m))) +vfloat16m2_t vslideup(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m4))) +vfloat16m4_t vslideup(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m4_m))) +vfloat16m4_t vslideup(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m8))) +vfloat16m8_t vslideup(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m8_m))) +vfloat16m8_t vslideup(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16mf2))) +vfloat16mf2_t vslideup(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16mf2_m))) +vfloat16mf2_t vslideup(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16mf4))) +vfloat16mf4_t vslideup(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16mf4_m))) +vfloat16mf4_t vslideup(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m1))) +vfloat16m1_t vslidedown(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m1_m))) +vfloat16m1_t vslidedown(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m2))) +vfloat16m2_t vslidedown(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m2_m))) +vfloat16m2_t vslidedown(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m4))) +vfloat16m4_t vslidedown(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m4_m))) +vfloat16m4_t vslidedown(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m8))) +vfloat16m8_t vslidedown(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m8_m))) +vfloat16m8_t vslidedown(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16mf2))) +vfloat16mf2_t vslidedown(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16mf2_m))) +vfloat16mf2_t vslidedown(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16mf4))) +vfloat16mf4_t vslidedown(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16mf4_m))) +vfloat16mf4_t vslidedown(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m1))) +vfloat16m1_t vfslide1up(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m1_m))) +vfloat16m1_t vfslide1up(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m2))) +vfloat16m2_t vfslide1up(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m2_m))) +vfloat16m2_t vfslide1up(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m4))) +vfloat16m4_t vfslide1up(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m4_m))) +vfloat16m4_t vfslide1up(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m8))) +vfloat16m8_t vfslide1up(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m8_m))) +vfloat16m8_t vfslide1up(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16mf2))) +vfloat16mf2_t vfslide1up(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16mf2_m))) +vfloat16mf2_t vfslide1up(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16mf4))) +vfloat16mf4_t vfslide1up(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16mf4_m))) +vfloat16mf4_t vfslide1up(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16m1))) +void vsuxei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16m1_m))) +void vsuxei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16m2))) +void vsuxei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16m2_m))) +void vsuxei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16m4))) +void vsuxei32(_Float16 * op0, vuint32m8_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16m4_m))) +void vsuxei32(vbool4_t op0, _Float16 * op1, vuint32m8_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16mf2))) +void vsuxei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16mf2_m))) +void vsuxei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16mf4))) +void vsuxei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16mf4_m))) +void vsuxei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m1))) +vfloat16m1_t vfslide1down(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m1_m))) +vfloat16m1_t vfslide1down(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m2))) +vfloat16m2_t vfslide1down(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m2_m))) +vfloat16m2_t vfslide1down(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m4))) +vfloat16m4_t vfslide1down(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m4_m))) +vfloat16m4_t vfslide1down(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m8))) +vfloat16m8_t vfslide1down(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m8_m))) +vfloat16m8_t vfslide1down(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16mf2))) +vfloat16mf2_t vfslide1down(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16mf2_m))) +vfloat16mf2_t vfslide1down(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16mf4))) +vfloat16mf4_t vfslide1down(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16mf4_m))) +vfloat16mf4_t vfslide1down(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m1))) +vfloat16m1_t vrgather(vfloat16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m1_m))) +vfloat16m1_t vrgather(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m2))) +vfloat16m2_t vrgather(vfloat16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m2_m))) +vfloat16m2_t vrgather(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m4))) +vfloat16m4_t vrgather(vfloat16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m4_m))) +vfloat16m4_t vrgather(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m8))) +vfloat16m8_t vrgather(vfloat16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m8_m))) +vfloat16m8_t vrgather(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16mf2))) +vfloat16mf2_t vrgather(vfloat16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16mf2_m))) +vfloat16mf2_t vrgather(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16mf4))) +vfloat16mf4_t vrgather(vfloat16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16mf4_m))) +vfloat16mf4_t vrgather(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m1))) +vfloat16m1_t vrgather(vfloat16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m1_m))) +vfloat16m1_t vrgather(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m2))) +vfloat16m2_t vrgather(vfloat16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m2_m))) +vfloat16m2_t vrgather(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m4))) +vfloat16m4_t vrgather(vfloat16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m4_m))) +vfloat16m4_t vrgather(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m8))) +vfloat16m8_t vrgather(vfloat16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m8_m))) +vfloat16m8_t vrgather(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16mf2))) +vfloat16mf2_t vrgather(vfloat16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16mf2_m))) +vfloat16mf2_t vrgather(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16mf4))) +vfloat16mf4_t vrgather(vfloat16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16mf4_m))) +vfloat16mf4_t vrgather(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m1))) +vfloat16m1_t vrgatherei16(vfloat16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m1_m))) +vfloat16m1_t vrgatherei16(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m2))) +vfloat16m2_t vrgatherei16(vfloat16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m2_m))) +vfloat16m2_t vrgatherei16(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m4))) +vfloat16m4_t vrgatherei16(vfloat16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m4_m))) +vfloat16m4_t vrgatherei16(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m8))) +vfloat16m8_t vrgatherei16(vfloat16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m8_m))) +vfloat16m8_t vrgatherei16(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16mf2))) +vfloat16mf2_t vrgatherei16(vfloat16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16mf2_m))) +vfloat16mf2_t vrgatherei16(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16mf4))) +vfloat16mf4_t vrgatherei16(vfloat16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16mf4_m))) +vfloat16mf4_t vrgatherei16(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f16m1))) +vfloat16m1_t vcompress(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f16m2))) +vfloat16m2_t vcompress(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f16m4))) +vfloat16m4_t vcompress(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f16m8))) +vfloat16m8_t vcompress(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f16mf2))) +vfloat16mf2_t vcompress(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f16mf4))) +vfloat16mf4_t vcompress(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16m1))) +void vsuxei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16m1_m))) +void vsuxei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16m2))) +void vsuxei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16m2_m))) +void vsuxei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16mf2))) +void vsuxei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16mf2_m))) +void vsuxei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16mf4))) +void vsuxei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16mf4_m))) +void vsuxei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m1))) +void vsoxei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m1_m))) +void vsoxei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m2))) +void vsoxei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m2_m))) +void vsoxei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m4))) +void vsoxei8(_Float16 * op0, vuint8m2_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m4_m))) +void vsoxei8(vbool4_t op0, _Float16 * op1, vuint8m2_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m8))) +void vsoxei8(_Float16 * op0, vuint8m4_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m8_m))) +void vsoxei8(vbool2_t op0, _Float16 * op1, vuint8m4_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16mf2))) +void vsoxei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16mf2_m))) +void vsoxei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16mf4))) +void vsoxei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16mf4_m))) +void vsoxei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m1))) +void vsoxei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m1_m))) +void vsoxei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m2))) +void vsoxei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m2_m))) +void vsoxei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m4))) +void vsoxei16(_Float16 * op0, vuint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m4_m))) +void vsoxei16(vbool4_t op0, _Float16 * op1, vuint16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m8))) +void vsoxei16(_Float16 * op0, vuint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m8_m))) +void vsoxei16(vbool2_t op0, _Float16 * op1, vuint16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16mf2))) +void vsoxei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16mf2_m))) +void vsoxei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16mf4))) +void vsoxei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16mf4_m))) +void vsoxei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16m1))) +void vsoxei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16m1_m))) +void vsoxei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16m2))) +void vsoxei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16m2_m))) +void vsoxei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16m4))) +void vsoxei32(_Float16 * op0, vuint32m8_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16m4_m))) +void vsoxei32(vbool4_t op0, _Float16 * op1, vuint32m8_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16mf2))) +void vsoxei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16mf2_m))) +void vsoxei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16mf4))) +void vsoxei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16mf4_m))) +void vsoxei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16m1))) +void vsoxei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16m1_m))) +void vsoxei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16m2))) +void vsoxei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16m2_m))) +void vsoxei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16mf2))) +void vsoxei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16mf2_m))) +void vsoxei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16mf4))) +void vsoxei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16mf4_m))) +void vsoxei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_f16m1_m))) +vfloat16m1_t vle16ff(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_f16m2_m))) +vfloat16m2_t vle16ff(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_f16m4_m))) +vfloat16m4_t vle16ff(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_f16m8_m))) +vfloat16m8_t vle16ff(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_f16mf2_m))) +vfloat16mf2_t vle16ff(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_f16mf4_m))) +vfloat16mf4_t vle16ff(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m1))) +vfloat16m1_t vfneg(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m1_m))) +vfloat16m1_t vfneg(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m2))) +vfloat16m2_t vfneg(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m2_m))) +vfloat16m2_t vfneg(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m4))) +vfloat16m4_t vfneg(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m4_m))) +vfloat16m4_t vfneg(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m8))) +vfloat16m8_t vfneg(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m8_m))) +vfloat16m8_t vfneg(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16mf2))) +vfloat16mf2_t vfneg(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16mf2_m))) +vfloat16mf2_t vfneg(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16mf4))) +vfloat16mf4_t vfneg(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16mf4_m))) +vfloat16mf4_t vfneg(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_f16m1_m))) +vfloat16m1_t vle16(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_f16m2_m))) +vfloat16m2_t vle16(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_f16m4_m))) +vfloat16m4_t vle16(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_f16m8_m))) +vfloat16m8_t vle16(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_f16mf2_m))) +vfloat16mf2_t vle16(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_f16mf4_m))) +vfloat16mf4_t vle16(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m1))) +void vse16(_Float16 * op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m1_m))) +void vse16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m2))) +void vse16(_Float16 * op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m2_m))) +void vse16(vbool8_t op0, _Float16 * op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m4))) +void vse16(_Float16 * op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m4_m))) +void vse16(vbool4_t op0, _Float16 * op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m8))) +void vse16(_Float16 * op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m8_m))) +void vse16(vbool2_t op0, _Float16 * op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16mf2))) +void vse16(_Float16 * op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16mf2_m))) +void vse16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16mf4))) +void vse16(_Float16 * op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16mf4_m))) +void vse16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_f16m1_m))) +vfloat16m1_t vlse16(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_f16m2_m))) +vfloat16m2_t vlse16(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_f16m4_m))) +vfloat16m4_t vlse16(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_f16m8_m))) +vfloat16m8_t vlse16(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_f16mf2_m))) +vfloat16mf2_t vlse16(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_f16mf4_m))) +vfloat16mf4_t vlse16(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m1))) +vuint16m1_t vfclass(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m1_m))) +vuint16m1_t vfclass(vbool16_t op0, vuint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m2))) +vuint16m2_t vfclass(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m2_m))) +vuint16m2_t vfclass(vbool8_t op0, vuint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m4))) +vuint16m4_t vfclass(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m4_m))) +vuint16m4_t vfclass(vbool4_t op0, vuint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m8))) +vuint16m8_t vfclass(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m8_m))) +vuint16m8_t vfclass(vbool2_t op0, vuint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16mf2))) +vuint16mf2_t vfclass(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16mf2_m))) +vuint16mf2_t vfclass(vbool32_t op0, vuint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16mf4))) +vuint16mf4_t vfclass(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16mf4_m))) +vuint16mf4_t vfclass(vbool64_t op0, vuint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m1))) +vfloat16m1_t vfcvt_f(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m1_m))) +vfloat16m1_t vfcvt_f(vbool16_t op0, vfloat16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m2))) +vfloat16m2_t vfcvt_f(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m2_m))) +vfloat16m2_t vfcvt_f(vbool8_t op0, vfloat16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m4))) +vfloat16m4_t vfcvt_f(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m4_m))) +vfloat16m4_t vfcvt_f(vbool4_t op0, vfloat16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m8))) +vfloat16m8_t vfcvt_f(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m8_m))) +vfloat16m8_t vfcvt_f(vbool2_t op0, vfloat16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16mf2))) +vfloat16mf2_t vfcvt_f(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16mf2_m))) +vfloat16mf2_t vfcvt_f(vbool32_t op0, vfloat16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16mf4))) +vfloat16mf4_t vfcvt_f(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16mf4_m))) +vfloat16mf4_t vfcvt_f(vbool64_t op0, vfloat16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m1))) +vfloat16m1_t vfcvt_f(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m1_m))) +vfloat16m1_t vfcvt_f(vbool16_t op0, vfloat16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m2))) +vfloat16m2_t vfcvt_f(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m2_m))) +vfloat16m2_t vfcvt_f(vbool8_t op0, vfloat16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m4))) +vfloat16m4_t vfcvt_f(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m4_m))) +vfloat16m4_t vfcvt_f(vbool4_t op0, vfloat16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m8))) +vfloat16m8_t vfcvt_f(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m8_m))) +vfloat16m8_t vfcvt_f(vbool2_t op0, vfloat16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16mf2))) +vfloat16mf2_t vfcvt_f(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16mf2_m))) +vfloat16mf2_t vfcvt_f(vbool32_t op0, vfloat16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16mf4))) +vfloat16mf4_t vfcvt_f(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16mf4_m))) +vfloat16mf4_t vfcvt_f(vbool64_t op0, vfloat16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m1))) +vint16m1_t vfcvt_rtz_x(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m1_m))) +vint16m1_t vfcvt_rtz_x(vbool16_t op0, vint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m2))) +vint16m2_t vfcvt_rtz_x(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m2_m))) +vint16m2_t vfcvt_rtz_x(vbool8_t op0, vint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m4))) +vint16m4_t vfcvt_rtz_x(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m4_m))) +vint16m4_t vfcvt_rtz_x(vbool4_t op0, vint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m8))) +vint16m8_t vfcvt_rtz_x(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m8_m))) +vint16m8_t vfcvt_rtz_x(vbool2_t op0, vint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16mf2))) +vint16mf2_t vfcvt_rtz_x(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16mf2_m))) +vint16mf2_t vfcvt_rtz_x(vbool32_t op0, vint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16mf4))) +vint16mf4_t vfcvt_rtz_x(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16mf4_m))) +vint16mf4_t vfcvt_rtz_x(vbool64_t op0, vint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m1))) +vuint16m1_t vfcvt_rtz_xu(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m1_m))) +vuint16m1_t vfcvt_rtz_xu(vbool16_t op0, vuint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m2))) +vuint16m2_t vfcvt_rtz_xu(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m2_m))) +vuint16m2_t vfcvt_rtz_xu(vbool8_t op0, vuint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m4))) +vuint16m4_t vfcvt_rtz_xu(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m4_m))) +vuint16m4_t vfcvt_rtz_xu(vbool4_t op0, vuint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m8))) +vuint16m8_t vfcvt_rtz_xu(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m8_m))) +vuint16m8_t vfcvt_rtz_xu(vbool2_t op0, vuint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf2))) +vuint16mf2_t vfcvt_rtz_xu(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf2_m))) +vuint16mf2_t vfcvt_rtz_xu(vbool32_t op0, vuint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf4))) +vuint16mf4_t vfcvt_rtz_xu(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf4_m))) +vuint16mf4_t vfcvt_rtz_xu(vbool64_t op0, vuint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m1))) +vint16m1_t vfcvt_x(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m1_m))) +vint16m1_t vfcvt_x(vbool16_t op0, vint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m2))) +vint16m2_t vfcvt_x(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m2_m))) +vint16m2_t vfcvt_x(vbool8_t op0, vint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m4))) +vint16m4_t vfcvt_x(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m4_m))) +vint16m4_t vfcvt_x(vbool4_t op0, vint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m8))) +vint16m8_t vfcvt_x(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m8_m))) +vint16m8_t vfcvt_x(vbool2_t op0, vint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16mf2))) +vint16mf2_t vfcvt_x(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16mf2_m))) +vint16mf2_t vfcvt_x(vbool32_t op0, vint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16mf4))) +vint16mf4_t vfcvt_x(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16mf4_m))) +vint16mf4_t vfcvt_x(vbool64_t op0, vint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m1))) +vuint16m1_t vfcvt_xu(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m1_m))) +vuint16m1_t vfcvt_xu(vbool16_t op0, vuint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m2))) +vuint16m2_t vfcvt_xu(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m2_m))) +vuint16m2_t vfcvt_xu(vbool8_t op0, vuint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m4))) +vuint16m4_t vfcvt_xu(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m4_m))) +vuint16m4_t vfcvt_xu(vbool4_t op0, vuint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m8))) +vuint16m8_t vfcvt_xu(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m8_m))) +vuint16m8_t vfcvt_xu(vbool2_t op0, vuint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16mf2))) +vuint16mf2_t vfcvt_xu(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16mf2_m))) +vuint16mf2_t vfcvt_xu(vbool32_t op0, vuint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16mf4))) +vuint16mf4_t vfcvt_xu(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16mf4_m))) +vuint16mf4_t vfcvt_xu(vbool64_t op0, vuint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16mf4))) +vfloat16mf4_t vfncvt_f(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16mf4_m))) +vfloat16mf4_t vfncvt_f(vbool64_t op0, vfloat16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16mf2))) +vfloat16mf2_t vfncvt_f(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16mf2_m))) +vfloat16mf2_t vfncvt_f(vbool32_t op0, vfloat16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16m1))) +vfloat16m1_t vfncvt_f(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16m1_m))) +vfloat16m1_t vfncvt_f(vbool16_t op0, vfloat16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16m2))) +vfloat16m2_t vfncvt_f(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16m2_m))) +vfloat16m2_t vfncvt_f(vbool8_t op0, vfloat16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16m4))) +vfloat16m4_t vfncvt_f(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16m4_m))) +vfloat16m4_t vfncvt_f(vbool4_t op0, vfloat16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16mf4))) +vfloat16mf4_t vfncvt_f(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16mf4_m))) +vfloat16mf4_t vfncvt_f(vbool64_t op0, vfloat16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16mf2))) +vfloat16mf2_t vfncvt_f(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16mf2_m))) +vfloat16mf2_t vfncvt_f(vbool32_t op0, vfloat16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16m1))) +vfloat16m1_t vfncvt_f(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16m1_m))) +vfloat16m1_t vfncvt_f(vbool16_t op0, vfloat16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16m2))) +vfloat16m2_t vfncvt_f(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16m2_m))) +vfloat16m2_t vfncvt_f(vbool8_t op0, vfloat16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16m4))) +vfloat16m4_t vfncvt_f(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16m4_m))) +vfloat16m4_t vfncvt_f(vbool4_t op0, vfloat16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8mf8))) +vint8mf8_t vfncvt_rtz_x(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8mf8_m))) +vint8mf8_t vfncvt_rtz_x(vbool64_t op0, vint8mf8_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8mf4))) +vint8mf4_t vfncvt_rtz_x(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8mf4_m))) +vint8mf4_t vfncvt_rtz_x(vbool32_t op0, vint8mf4_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8mf2))) +vint8mf2_t vfncvt_rtz_x(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8mf2_m))) +vint8mf2_t vfncvt_rtz_x(vbool16_t op0, vint8mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8m1))) +vint8m1_t vfncvt_rtz_x(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8m1_m))) +vint8m1_t vfncvt_rtz_x(vbool8_t op0, vint8m1_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8m2))) +vint8m2_t vfncvt_rtz_x(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8m2_m))) +vint8m2_t vfncvt_rtz_x(vbool4_t op0, vint8m2_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8m4))) +vint8m4_t vfncvt_rtz_x(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8m4_m))) +vint8m4_t vfncvt_rtz_x(vbool2_t op0, vint8m4_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf8))) +vuint8mf8_t vfncvt_rtz_xu(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf8_m))) +vuint8mf8_t vfncvt_rtz_xu(vbool64_t op0, vuint8mf8_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf4))) +vuint8mf4_t vfncvt_rtz_xu(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf4_m))) +vuint8mf4_t vfncvt_rtz_xu(vbool32_t op0, vuint8mf4_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf2))) +vuint8mf2_t vfncvt_rtz_xu(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf2_m))) +vuint8mf2_t vfncvt_rtz_xu(vbool16_t op0, vuint8mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8m1))) +vuint8m1_t vfncvt_rtz_xu(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8m1_m))) +vuint8m1_t vfncvt_rtz_xu(vbool8_t op0, vuint8m1_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8m2))) +vuint8m2_t vfncvt_rtz_xu(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8m2_m))) +vuint8m2_t vfncvt_rtz_xu(vbool4_t op0, vuint8m2_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8m4))) +vuint8m4_t vfncvt_rtz_xu(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8m4_m))) +vuint8m4_t vfncvt_rtz_xu(vbool2_t op0, vuint8m4_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8mf8))) +vint8mf8_t vfncvt_x(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8mf8_m))) +vint8mf8_t vfncvt_x(vbool64_t op0, vint8mf8_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8mf4))) +vint8mf4_t vfncvt_x(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8mf4_m))) +vint8mf4_t vfncvt_x(vbool32_t op0, vint8mf4_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8mf2))) +vint8mf2_t vfncvt_x(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8mf2_m))) +vint8mf2_t vfncvt_x(vbool16_t op0, vint8mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8m1))) +vint8m1_t vfncvt_x(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8m1_m))) +vint8m1_t vfncvt_x(vbool8_t op0, vint8m1_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8m2))) +vint8m2_t vfncvt_x(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8m2_m))) +vint8m2_t vfncvt_x(vbool4_t op0, vint8m2_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8m4))) +vint8m4_t vfncvt_x(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8m4_m))) +vint8m4_t vfncvt_x(vbool2_t op0, vint8m4_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8mf8))) +vuint8mf8_t vfncvt_xu(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8mf8_m))) +vuint8mf8_t vfncvt_xu(vbool64_t op0, vuint8mf8_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8mf4))) +vuint8mf4_t vfncvt_xu(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8mf4_m))) +vuint8mf4_t vfncvt_xu(vbool32_t op0, vuint8mf4_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8mf2))) +vuint8mf2_t vfncvt_xu(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8mf2_m))) +vuint8mf2_t vfncvt_xu(vbool16_t op0, vuint8mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8m1))) +vuint8m1_t vfncvt_xu(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8m1_m))) +vuint8m1_t vfncvt_xu(vbool8_t op0, vuint8m1_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8m2))) +vuint8m2_t vfncvt_xu(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8m2_m))) +vuint8m2_t vfncvt_xu(vbool4_t op0, vuint8m2_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8m4))) +vuint8m4_t vfncvt_xu(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8m4_m))) +vuint8m4_t vfncvt_xu(vbool2_t op0, vuint8m4_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m1))) +vfloat16m1_t vfrec7(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m1_m))) +vfloat16m1_t vfrec7(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m2))) +vfloat16m2_t vfrec7(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m2_m))) +vfloat16m2_t vfrec7(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m4))) +vfloat16m4_t vfrec7(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m4_m))) +vfloat16m4_t vfrec7(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m8))) +vfloat16m8_t vfrec7(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m8_m))) +vfloat16m8_t vfrec7(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16mf2))) +vfloat16mf2_t vfrec7(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16mf2_m))) +vfloat16mf2_t vfrec7(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16mf4))) +vfloat16mf4_t vfrec7(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16mf4_m))) +vfloat16mf4_t vfrec7(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m1))) +vfloat16m1_t vfrsqrt7(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m1_m))) +vfloat16m1_t vfrsqrt7(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m2))) +vfloat16m2_t vfrsqrt7(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m2_m))) +vfloat16m2_t vfrsqrt7(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m4))) +vfloat16m4_t vfrsqrt7(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m4_m))) +vfloat16m4_t vfrsqrt7(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m8))) +vfloat16m8_t vfrsqrt7(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m8_m))) +vfloat16m8_t vfrsqrt7(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16mf2))) +vfloat16mf2_t vfrsqrt7(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16mf2_m))) +vfloat16mf2_t vfrsqrt7(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16mf4))) +vfloat16mf4_t vfrsqrt7(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16mf4_m))) +vfloat16mf4_t vfrsqrt7(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m1))) +vfloat16m1_t vfsqrt(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m1_m))) +vfloat16m1_t vfsqrt(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m2))) +vfloat16m2_t vfsqrt(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m2_m))) +vfloat16m2_t vfsqrt(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m4))) +vfloat16m4_t vfsqrt(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m4_m))) +vfloat16m4_t vfsqrt(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m8))) +vfloat16m8_t vfsqrt(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m8_m))) +vfloat16m8_t vfsqrt(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16mf2))) +vfloat16mf2_t vfsqrt(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16mf2_m))) +vfloat16mf2_t vfsqrt(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16mf4))) +vfloat16mf4_t vfsqrt(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16mf4_m))) +vfloat16mf4_t vfsqrt(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16mf4))) +vfloat16mf4_t vfwcvt_f(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16mf4_m))) +vfloat16mf4_t vfwcvt_f(vbool64_t op0, vfloat16mf4_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16mf2))) +vfloat16mf2_t vfwcvt_f(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16mf2_m))) +vfloat16mf2_t vfwcvt_f(vbool32_t op0, vfloat16mf2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m1))) +vfloat16m1_t vfwcvt_f(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m1_m))) +vfloat16m1_t vfwcvt_f(vbool16_t op0, vfloat16m1_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m2))) +vfloat16m2_t vfwcvt_f(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m2_m))) +vfloat16m2_t vfwcvt_f(vbool8_t op0, vfloat16m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m4))) +vfloat16m4_t vfwcvt_f(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m4_m))) +vfloat16m4_t vfwcvt_f(vbool4_t op0, vfloat16m4_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m8))) +vfloat16m8_t vfwcvt_f(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m8_m))) +vfloat16m8_t vfwcvt_f(vbool2_t op0, vfloat16m8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16mf4))) +vfloat16mf4_t vfwcvt_f(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16mf4_m))) +vfloat16mf4_t vfwcvt_f(vbool64_t op0, vfloat16mf4_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16mf2))) +vfloat16mf2_t vfwcvt_f(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16mf2_m))) +vfloat16mf2_t vfwcvt_f(vbool32_t op0, vfloat16mf2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m1))) +vfloat16m1_t vfwcvt_f(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m1_m))) +vfloat16m1_t vfwcvt_f(vbool16_t op0, vfloat16m1_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m2))) +vfloat16m2_t vfwcvt_f(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m2_m))) +vfloat16m2_t vfwcvt_f(vbool8_t op0, vfloat16m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m4))) +vfloat16m4_t vfwcvt_f(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m4_m))) +vfloat16m4_t vfwcvt_f(vbool4_t op0, vfloat16m4_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m8))) +vfloat16m8_t vfwcvt_f(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m8_m))) +vfloat16m8_t vfwcvt_f(vbool2_t op0, vfloat16m8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32mf2))) +vint32mf2_t vfwcvt_rtz_x(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32mf2_m))) +vint32mf2_t vfwcvt_rtz_x(vbool64_t op0, vint32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m1))) +vint32m1_t vfwcvt_rtz_x(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m1_m))) +vint32m1_t vfwcvt_rtz_x(vbool32_t op0, vint32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m2))) +vint32m2_t vfwcvt_rtz_x(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m2_m))) +vint32m2_t vfwcvt_rtz_x(vbool16_t op0, vint32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m4))) +vint32m4_t vfwcvt_rtz_x(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m4_m))) +vint32m4_t vfwcvt_rtz_x(vbool8_t op0, vint32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m8))) +vint32m8_t vfwcvt_rtz_x(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m8_m))) +vint32m8_t vfwcvt_rtz_x(vbool4_t op0, vint32m8_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32mf2))) +vuint32mf2_t vfwcvt_rtz_xu(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32mf2_m))) +vuint32mf2_t vfwcvt_rtz_xu(vbool64_t op0, vuint32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m1))) +vuint32m1_t vfwcvt_rtz_xu(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m1_m))) +vuint32m1_t vfwcvt_rtz_xu(vbool32_t op0, vuint32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m2))) +vuint32m2_t vfwcvt_rtz_xu(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m2_m))) +vuint32m2_t vfwcvt_rtz_xu(vbool16_t op0, vuint32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m4))) +vuint32m4_t vfwcvt_rtz_xu(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m4_m))) +vuint32m4_t vfwcvt_rtz_xu(vbool8_t op0, vuint32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m8))) +vuint32m8_t vfwcvt_rtz_xu(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m8_m))) +vuint32m8_t vfwcvt_rtz_xu(vbool4_t op0, vuint32m8_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32mf2))) +vint32mf2_t vfwcvt_x(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32mf2_m))) +vint32mf2_t vfwcvt_x(vbool64_t op0, vint32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m1))) +vint32m1_t vfwcvt_x(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m1_m))) +vint32m1_t vfwcvt_x(vbool32_t op0, vint32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m2))) +vint32m2_t vfwcvt_x(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m2_m))) +vint32m2_t vfwcvt_x(vbool16_t op0, vint32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m4))) +vint32m4_t vfwcvt_x(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m4_m))) +vint32m4_t vfwcvt_x(vbool8_t op0, vint32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m8))) +vint32m8_t vfwcvt_x(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m8_m))) +vint32m8_t vfwcvt_x(vbool4_t op0, vint32m8_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32mf2))) +vuint32mf2_t vfwcvt_xu(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32mf2_m))) +vuint32mf2_t vfwcvt_xu(vbool64_t op0, vuint32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m1))) +vuint32m1_t vfwcvt_xu(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m1_m))) +vuint32m1_t vfwcvt_xu(vbool32_t op0, vuint32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m2))) +vuint32m2_t vfwcvt_xu(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m2_m))) +vuint32m2_t vfwcvt_xu(vbool16_t op0, vuint32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m4))) +vuint32m4_t vfwcvt_xu(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m4_m))) +vuint32m4_t vfwcvt_xu(vbool8_t op0, vuint32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m8))) +vuint32m8_t vfwcvt_xu(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m8_m))) +vuint32m8_t vfwcvt_xu(vbool4_t op0, vuint32m8_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf2_f16m1))) +vfloat16m1_t vlmul_ext_f16m1(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf4_f16m1))) +vfloat16m1_t vlmul_ext_f16m1(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf4_f16mf2))) +vfloat16mf2_t vlmul_ext_f16mf2(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16m1_f16m2))) +vfloat16m2_t vlmul_ext_f16m2(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf2_f16m2))) +vfloat16m2_t vlmul_ext_f16m2(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf4_f16m2))) +vfloat16m2_t vlmul_ext_f16m2(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16m1_f16m4))) +vfloat16m4_t vlmul_ext_f16m4(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16m2_f16m4))) +vfloat16m4_t vlmul_ext_f16m4(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf2_f16m4))) +vfloat16m4_t vlmul_ext_f16m4(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf4_f16m4))) +vfloat16m4_t vlmul_ext_f16m4(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16m1_f16m8))) +vfloat16m8_t vlmul_ext_f16m8(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16m2_f16m8))) +vfloat16m8_t vlmul_ext_f16m8(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16m4_f16m8))) +vfloat16m8_t vlmul_ext_f16m8(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf2_f16m8))) +vfloat16m8_t vlmul_ext_f16m8(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf4_f16m8))) +vfloat16m8_t vlmul_ext_f16m8(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m1_f16mf2))) +vfloat16mf2_t vlmul_trunc_f16mf2(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m2_f16mf2))) +vfloat16mf2_t vlmul_trunc_f16mf2(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m4_f16mf2))) +vfloat16mf2_t vlmul_trunc_f16mf2(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m8_f16mf2))) +vfloat16mf2_t vlmul_trunc_f16mf2(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m1_f16mf4))) +vfloat16mf4_t vlmul_trunc_f16mf4(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m2_f16mf4))) +vfloat16mf4_t vlmul_trunc_f16mf4(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m4_f16mf4))) +vfloat16mf4_t vlmul_trunc_f16mf4(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m8_f16mf4))) +vfloat16mf4_t vlmul_trunc_f16mf4(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16mf2_f16mf4))) +vfloat16mf4_t vlmul_trunc_f16mf4(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m2_f16m1))) +vfloat16m1_t vlmul_trunc_f16m1(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m4_f16m1))) +vfloat16m1_t vlmul_trunc_f16m1(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m8_f16m1))) +vfloat16m1_t vlmul_trunc_f16m1(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m4_f16m2))) +vfloat16m2_t vlmul_trunc_f16m2(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m8_f16m2))) +vfloat16m2_t vlmul_trunc_f16m2(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m8_f16m4))) +vfloat16m4_t vlmul_trunc_f16m4(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m1_f16m1))) +vfloat16m1_t vreinterpret_f16m1(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m2_f16m2))) +vfloat16m2_t vreinterpret_f16m2(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m4_f16m4))) +vfloat16m4_t vreinterpret_f16m4(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m8_f16m8))) +vfloat16m8_t vreinterpret_f16m8(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf2_f16mf2))) +vfloat16mf2_t vreinterpret_f16mf2(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf4_f16mf4))) +vfloat16mf4_t vreinterpret_f16mf4(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m1_f16m1))) +vfloat16m1_t vreinterpret_f16m1(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m2_f16m2))) +vfloat16m2_t vreinterpret_f16m2(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m4_f16m4))) +vfloat16m4_t vreinterpret_f16m4(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m8_f16m8))) +vfloat16m8_t vreinterpret_f16m8(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf2_f16mf2))) +vfloat16mf2_t vreinterpret_f16mf2(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf4_f16mf4))) +vfloat16mf4_t vreinterpret_f16mf4(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m1_i16m1))) +vint16m1_t vreinterpret_i16m1(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m2_i16m2))) +vint16m2_t vreinterpret_i16m2(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m4_i16m4))) +vint16m4_t vreinterpret_i16m4(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m8_i16m8))) +vint16m8_t vreinterpret_i16m8(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16mf2_i16mf2))) +vint16mf2_t vreinterpret_i16mf2(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16mf4_i16mf4))) +vint16mf4_t vreinterpret_i16mf4(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m1_u16m1))) +vuint16m1_t vreinterpret_u16m1(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m2_u16m2))) +vuint16m2_t vreinterpret_u16m2(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m4_u16m4))) +vuint16m4_t vreinterpret_u16m4(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m8_u16m8))) +vuint16m8_t vreinterpret_u16m8(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16mf2_u16mf2))) +vuint16mf2_t vreinterpret_u16mf2(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16mf4_u16mf4))) +vuint16mf4_t vreinterpret_u16mf4(vfloat16mf4_t op0); + +#endif + +#if defined(__riscv_f) && defined(__riscv_zfh) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32mf2))) +vfloat32mf2_t vfwadd_vv(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32mf2_m))) +vfloat32mf2_t vfwadd_vv(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m1))) +vfloat32m1_t vfwadd_vv(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m1_m))) +vfloat32m1_t vfwadd_vv(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m2))) +vfloat32m2_t vfwadd_vv(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m2_m))) +vfloat32m2_t vfwadd_vv(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m4))) +vfloat32m4_t vfwadd_vv(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m4_m))) +vfloat32m4_t vfwadd_vv(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m8))) +vfloat32m8_t vfwadd_vv(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m8_m))) +vfloat32m8_t vfwadd_vv(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32mf2))) +vfloat32mf2_t vfwadd_vf(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32mf2_m))) +vfloat32mf2_t vfwadd_vf(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m1))) +vfloat32m1_t vfwadd_vf(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m1_m))) +vfloat32m1_t vfwadd_vf(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m2))) +vfloat32m2_t vfwadd_vf(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m2_m))) +vfloat32m2_t vfwadd_vf(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m4))) +vfloat32m4_t vfwadd_vf(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m4_m))) +vfloat32m4_t vfwadd_vf(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m8))) +vfloat32m8_t vfwadd_vf(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m8_m))) +vfloat32m8_t vfwadd_vf(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32mf2))) +vfloat32mf2_t vfwsub_vv(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32mf2_m))) +vfloat32mf2_t vfwsub_vv(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m1))) +vfloat32m1_t vfwsub_vv(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m1_m))) +vfloat32m1_t vfwsub_vv(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m2))) +vfloat32m2_t vfwsub_vv(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m2_m))) +vfloat32m2_t vfwsub_vv(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m4))) +vfloat32m4_t vfwsub_vv(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m4_m))) +vfloat32m4_t vfwsub_vv(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m8))) +vfloat32m8_t vfwsub_vv(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m8_m))) +vfloat32m8_t vfwsub_vv(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32mf2))) +vfloat32mf2_t vfwsub_vf(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32mf2_m))) +vfloat32mf2_t vfwsub_vf(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m1))) +vfloat32m1_t vfwsub_vf(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m1_m))) +vfloat32m1_t vfwsub_vf(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m2))) +vfloat32m2_t vfwsub_vf(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m2_m))) +vfloat32m2_t vfwsub_vf(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m4))) +vfloat32m4_t vfwsub_vf(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m4_m))) +vfloat32m4_t vfwsub_vf(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m8))) +vfloat32m8_t vfwsub_vf(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m8_m))) +vfloat32m8_t vfwsub_vf(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32mf2))) +vfloat32mf2_t vfwadd_wv(vfloat32mf2_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32mf2_m))) +vfloat32mf2_t vfwadd_wv(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m1))) +vfloat32m1_t vfwadd_wv(vfloat32m1_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m1_m))) +vfloat32m1_t vfwadd_wv(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m2))) +vfloat32m2_t vfwadd_wv(vfloat32m2_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m2_m))) +vfloat32m2_t vfwadd_wv(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m4))) +vfloat32m4_t vfwadd_wv(vfloat32m4_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m4_m))) +vfloat32m4_t vfwadd_wv(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m8))) +vfloat32m8_t vfwadd_wv(vfloat32m8_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m8_m))) +vfloat32m8_t vfwadd_wv(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32mf2))) +vfloat32mf2_t vfwadd_wf(vfloat32mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32mf2_m))) +vfloat32mf2_t vfwadd_wf(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m1))) +vfloat32m1_t vfwadd_wf(vfloat32m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m1_m))) +vfloat32m1_t vfwadd_wf(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m2))) +vfloat32m2_t vfwadd_wf(vfloat32m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m2_m))) +vfloat32m2_t vfwadd_wf(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m4))) +vfloat32m4_t vfwadd_wf(vfloat32m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m4_m))) +vfloat32m4_t vfwadd_wf(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m8))) +vfloat32m8_t vfwadd_wf(vfloat32m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m8_m))) +vfloat32m8_t vfwadd_wf(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32mf2))) +vfloat32mf2_t vfwsub_wv(vfloat32mf2_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32mf2_m))) +vfloat32mf2_t vfwsub_wv(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m1))) +vfloat32m1_t vfwsub_wv(vfloat32m1_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m1_m))) +vfloat32m1_t vfwsub_wv(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m2))) +vfloat32m2_t vfwsub_wv(vfloat32m2_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m2_m))) +vfloat32m2_t vfwsub_wv(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m4))) +vfloat32m4_t vfwsub_wv(vfloat32m4_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m4_m))) +vfloat32m4_t vfwsub_wv(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m8))) +vfloat32m8_t vfwsub_wv(vfloat32m8_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m8_m))) +vfloat32m8_t vfwsub_wv(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32mf2))) +vfloat32mf2_t vfwsub_wf(vfloat32mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32mf2_m))) +vfloat32mf2_t vfwsub_wf(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m1))) +vfloat32m1_t vfwsub_wf(vfloat32m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m1_m))) +vfloat32m1_t vfwsub_wf(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m2))) +vfloat32m2_t vfwsub_wf(vfloat32m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m2_m))) +vfloat32m2_t vfwsub_wf(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m4))) +vfloat32m4_t vfwsub_wf(vfloat32m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m4_m))) +vfloat32m4_t vfwsub_wf(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m8))) +vfloat32m8_t vfwsub_wf(vfloat32m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m8_m))) +vfloat32m8_t vfwsub_wf(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32mf2))) +vfloat32mf2_t vfwmul(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32mf2_m))) +vfloat32mf2_t vfwmul(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m1))) +vfloat32m1_t vfwmul(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m1_m))) +vfloat32m1_t vfwmul(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m2))) +vfloat32m2_t vfwmul(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m2_m))) +vfloat32m2_t vfwmul(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m4))) +vfloat32m4_t vfwmul(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m4_m))) +vfloat32m4_t vfwmul(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m8))) +vfloat32m8_t vfwmul(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m8_m))) +vfloat32m8_t vfwmul(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32mf2))) +vfloat32mf2_t vfwmul(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32mf2_m))) +vfloat32mf2_t vfwmul(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m1))) +vfloat32m1_t vfwmul(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m1_m))) +vfloat32m1_t vfwmul(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m2))) +vfloat32m2_t vfwmul(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m2_m))) +vfloat32m2_t vfwmul(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m4))) +vfloat32m4_t vfwmul(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m4_m))) +vfloat32m4_t vfwmul(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m8))) +vfloat32m8_t vfwmul(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m8_m))) +vfloat32m8_t vfwmul(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32mf2))) +vfloat32mf2_t vfwmacc(vfloat32mf2_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32mf2_m))) +vfloat32mf2_t vfwmacc(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m1))) +vfloat32m1_t vfwmacc(vfloat32m1_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m1_m))) +vfloat32m1_t vfwmacc(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m2))) +vfloat32m2_t vfwmacc(vfloat32m2_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m2_m))) +vfloat32m2_t vfwmacc(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m4))) +vfloat32m4_t vfwmacc(vfloat32m4_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m4_m))) +vfloat32m4_t vfwmacc(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m8))) +vfloat32m8_t vfwmacc(vfloat32m8_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m8_m))) +vfloat32m8_t vfwmacc(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32mf2))) +vfloat32mf2_t vfwmacc(vfloat32mf2_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32mf2_m))) +vfloat32mf2_t vfwmacc(vbool64_t op0, vfloat32mf2_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m1))) +vfloat32m1_t vfwmacc(vfloat32m1_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m1_m))) +vfloat32m1_t vfwmacc(vbool32_t op0, vfloat32m1_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m2))) +vfloat32m2_t vfwmacc(vfloat32m2_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m2_m))) +vfloat32m2_t vfwmacc(vbool16_t op0, vfloat32m2_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m4))) +vfloat32m4_t vfwmacc(vfloat32m4_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m4_m))) +vfloat32m4_t vfwmacc(vbool8_t op0, vfloat32m4_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m8))) +vfloat32m8_t vfwmacc(vfloat32m8_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m8_m))) +vfloat32m8_t vfwmacc(vbool4_t op0, vfloat32m8_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32mf2))) +vfloat32mf2_t vfwnmacc(vfloat32mf2_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32mf2_m))) +vfloat32mf2_t vfwnmacc(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m1))) +vfloat32m1_t vfwnmacc(vfloat32m1_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m1_m))) +vfloat32m1_t vfwnmacc(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m2))) +vfloat32m2_t vfwnmacc(vfloat32m2_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m2_m))) +vfloat32m2_t vfwnmacc(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m4))) +vfloat32m4_t vfwnmacc(vfloat32m4_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m4_m))) +vfloat32m4_t vfwnmacc(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m8))) +vfloat32m8_t vfwnmacc(vfloat32m8_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m8_m))) +vfloat32m8_t vfwnmacc(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32mf2))) +vfloat32mf2_t vfwnmacc(vfloat32mf2_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32mf2_m))) +vfloat32mf2_t vfwnmacc(vbool64_t op0, vfloat32mf2_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m1))) +vfloat32m1_t vfwnmacc(vfloat32m1_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m1_m))) +vfloat32m1_t vfwnmacc(vbool32_t op0, vfloat32m1_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m2))) +vfloat32m2_t vfwnmacc(vfloat32m2_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m2_m))) +vfloat32m2_t vfwnmacc(vbool16_t op0, vfloat32m2_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m4))) +vfloat32m4_t vfwnmacc(vfloat32m4_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m4_m))) +vfloat32m4_t vfwnmacc(vbool8_t op0, vfloat32m4_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m8))) +vfloat32m8_t vfwnmacc(vfloat32m8_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m8_m))) +vfloat32m8_t vfwnmacc(vbool4_t op0, vfloat32m8_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32mf2))) +vfloat32mf2_t vfwmsac(vfloat32mf2_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32mf2_m))) +vfloat32mf2_t vfwmsac(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m1))) +vfloat32m1_t vfwmsac(vfloat32m1_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m1_m))) +vfloat32m1_t vfwmsac(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m2))) +vfloat32m2_t vfwmsac(vfloat32m2_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m2_m))) +vfloat32m2_t vfwmsac(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m4))) +vfloat32m4_t vfwmsac(vfloat32m4_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m4_m))) +vfloat32m4_t vfwmsac(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m8))) +vfloat32m8_t vfwmsac(vfloat32m8_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m8_m))) +vfloat32m8_t vfwmsac(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32mf2))) +vfloat32mf2_t vfwmsac(vfloat32mf2_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32mf2_m))) +vfloat32mf2_t vfwmsac(vbool64_t op0, vfloat32mf2_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m1))) +vfloat32m1_t vfwmsac(vfloat32m1_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m1_m))) +vfloat32m1_t vfwmsac(vbool32_t op0, vfloat32m1_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m2))) +vfloat32m2_t vfwmsac(vfloat32m2_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m2_m))) +vfloat32m2_t vfwmsac(vbool16_t op0, vfloat32m2_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m4))) +vfloat32m4_t vfwmsac(vfloat32m4_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m4_m))) +vfloat32m4_t vfwmsac(vbool8_t op0, vfloat32m4_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m8))) +vfloat32m8_t vfwmsac(vfloat32m8_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m8_m))) +vfloat32m8_t vfwmsac(vbool4_t op0, vfloat32m8_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32mf2))) +vfloat32mf2_t vfwnmsac(vfloat32mf2_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32mf2_m))) +vfloat32mf2_t vfwnmsac(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m1))) +vfloat32m1_t vfwnmsac(vfloat32m1_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m1_m))) +vfloat32m1_t vfwnmsac(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m2))) +vfloat32m2_t vfwnmsac(vfloat32m2_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m2_m))) +vfloat32m2_t vfwnmsac(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m4))) +vfloat32m4_t vfwnmsac(vfloat32m4_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m4_m))) +vfloat32m4_t vfwnmsac(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m8))) +vfloat32m8_t vfwnmsac(vfloat32m8_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m8_m))) +vfloat32m8_t vfwnmsac(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32mf2))) +vfloat32mf2_t vfwnmsac(vfloat32mf2_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32mf2_m))) +vfloat32mf2_t vfwnmsac(vbool64_t op0, vfloat32mf2_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m1))) +vfloat32m1_t vfwnmsac(vfloat32m1_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m1_m))) +vfloat32m1_t vfwnmsac(vbool32_t op0, vfloat32m1_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m2))) +vfloat32m2_t vfwnmsac(vfloat32m2_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m2_m))) +vfloat32m2_t vfwnmsac(vbool16_t op0, vfloat32m2_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m4))) +vfloat32m4_t vfwnmsac(vfloat32m4_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m4_m))) +vfloat32m4_t vfwnmsac(vbool8_t op0, vfloat32m4_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m8))) +vfloat32m8_t vfwnmsac(vfloat32m8_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m8_m))) +vfloat32m8_t vfwnmsac(vbool4_t op0, vfloat32m8_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m1_f32m1))) +vfloat32m1_t vfwredsum(vfloat32m1_t op0, vfloat16m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m1_f32m1_m))) +vfloat32m1_t vfwredsum(vbool16_t op0, vfloat32m1_t op1, vfloat16m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m2_f32m1))) +vfloat32m1_t vfwredsum(vfloat32m1_t op0, vfloat16m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m2_f32m1_m))) +vfloat32m1_t vfwredsum(vbool8_t op0, vfloat32m1_t op1, vfloat16m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m4_f32m1))) +vfloat32m1_t vfwredsum(vfloat32m1_t op0, vfloat16m4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m4_f32m1_m))) +vfloat32m1_t vfwredsum(vbool4_t op0, vfloat32m1_t op1, vfloat16m4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m8_f32m1))) +vfloat32m1_t vfwredsum(vfloat32m1_t op0, vfloat16m8_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m8_f32m1_m))) +vfloat32m1_t vfwredsum(vbool2_t op0, vfloat32m1_t op1, vfloat16m8_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16mf2_f32m1))) +vfloat32m1_t vfwredsum(vfloat32m1_t op0, vfloat16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16mf2_f32m1_m))) +vfloat32m1_t vfwredsum(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16mf4_f32m1))) +vfloat32m1_t vfwredsum(vfloat32m1_t op0, vfloat16mf4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16mf4_f32m1_m))) +vfloat32m1_t vfwredsum(vbool64_t op0, vfloat32m1_t op1, vfloat16mf4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m1_f32m1))) +vfloat32m1_t vfwredosum(vfloat32m1_t op0, vfloat16m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m1_f32m1_m))) +vfloat32m1_t vfwredosum(vbool16_t op0, vfloat32m1_t op1, vfloat16m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m2_f32m1))) +vfloat32m1_t vfwredosum(vfloat32m1_t op0, vfloat16m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m2_f32m1_m))) +vfloat32m1_t vfwredosum(vbool8_t op0, vfloat32m1_t op1, vfloat16m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m4_f32m1))) +vfloat32m1_t vfwredosum(vfloat32m1_t op0, vfloat16m4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m4_f32m1_m))) +vfloat32m1_t vfwredosum(vbool4_t op0, vfloat32m1_t op1, vfloat16m4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m8_f32m1))) +vfloat32m1_t vfwredosum(vfloat32m1_t op0, vfloat16m8_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m8_f32m1_m))) +vfloat32m1_t vfwredosum(vbool2_t op0, vfloat32m1_t op1, vfloat16m8_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16mf2_f32m1))) +vfloat32m1_t vfwredosum(vfloat32m1_t op0, vfloat16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16mf2_f32m1_m))) +vfloat32m1_t vfwredosum(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16mf4_f32m1))) +vfloat32m1_t vfwredosum(vfloat32m1_t op0, vfloat16mf4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16mf4_f32m1_m))) +vfloat32m1_t vfwredosum(vbool64_t op0, vfloat32m1_t op1, vfloat16mf4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16mf4))) +vfloat16mf4_t vfncvt_f(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16mf4_m))) +vfloat16mf4_t vfncvt_f(vbool64_t op0, vfloat16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16mf2))) +vfloat16mf2_t vfncvt_f(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16mf2_m))) +vfloat16mf2_t vfncvt_f(vbool32_t op0, vfloat16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16m1))) +vfloat16m1_t vfncvt_f(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16m1_m))) +vfloat16m1_t vfncvt_f(vbool16_t op0, vfloat16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16m2))) +vfloat16m2_t vfncvt_f(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16m2_m))) +vfloat16m2_t vfncvt_f(vbool8_t op0, vfloat16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16m4))) +vfloat16m4_t vfncvt_f(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16m4_m))) +vfloat16m4_t vfncvt_f(vbool4_t op0, vfloat16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16mf4))) +vfloat16mf4_t vfncvt_rod_f(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16mf4_m))) +vfloat16mf4_t vfncvt_rod_f(vbool64_t op0, vfloat16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16mf2))) +vfloat16mf2_t vfncvt_rod_f(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16mf2_m))) +vfloat16mf2_t vfncvt_rod_f(vbool32_t op0, vfloat16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16m1))) +vfloat16m1_t vfncvt_rod_f(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16m1_m))) +vfloat16m1_t vfncvt_rod_f(vbool16_t op0, vfloat16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16m2))) +vfloat16m2_t vfncvt_rod_f(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16m2_m))) +vfloat16m2_t vfncvt_rod_f(vbool8_t op0, vfloat16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16m4))) +vfloat16m4_t vfncvt_rod_f(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16m4_m))) +vfloat16m4_t vfncvt_rod_f(vbool4_t op0, vfloat16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32mf2))) +vfloat32mf2_t vfwcvt_f(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32mf2_m))) +vfloat32mf2_t vfwcvt_f(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m1))) +vfloat32m1_t vfwcvt_f(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m1_m))) +vfloat32m1_t vfwcvt_f(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m2))) +vfloat32m2_t vfwcvt_f(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m2_m))) +vfloat32m2_t vfwcvt_f(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m4))) +vfloat32m4_t vfwcvt_f(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m4_m))) +vfloat32m4_t vfwcvt_f(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m8))) +vfloat32m8_t vfwcvt_f(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m8_m))) +vfloat32m8_t vfwcvt_f(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, size_t op3); + +#endif + +#if defined(__riscv_zvamo) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m1))) +vint32m1_t vamoswapei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m1_m))) +vint32m1_t vamoswapei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m2))) +vint32m2_t vamoswapei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m2_m))) +vint32m2_t vamoswapei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m4))) +vint32m4_t vamoswapei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m4_m))) +vint32m4_t vamoswapei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m8))) +vint32m8_t vamoswapei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m8_m))) +vint32m8_t vamoswapei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32mf2))) +vint32mf2_t vamoswapei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32mf2_m))) +vint32mf2_t vamoswapei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m1))) +vuint32m1_t vamoswapei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m1_m))) +vuint32m1_t vamoswapei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m2))) +vuint32m2_t vamoswapei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m2_m))) +vuint32m2_t vamoswapei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m4))) +vuint32m4_t vamoswapei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m4_m))) +vuint32m4_t vamoswapei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m8))) +vuint32m8_t vamoswapei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m8_m))) +vuint32m8_t vamoswapei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32mf2))) +vuint32mf2_t vamoswapei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32mf2_m))) +vuint32mf2_t vamoswapei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m1))) +vint32m1_t vamoswapei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m1_m))) +vint32m1_t vamoswapei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m2))) +vint32m2_t vamoswapei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m2_m))) +vint32m2_t vamoswapei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m4))) +vint32m4_t vamoswapei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m4_m))) +vint32m4_t vamoswapei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m8))) +vint32m8_t vamoswapei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m8_m))) +vint32m8_t vamoswapei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32mf2))) +vint32mf2_t vamoswapei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32mf2_m))) +vint32mf2_t vamoswapei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m1))) +vuint32m1_t vamoswapei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m1_m))) +vuint32m1_t vamoswapei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m2))) +vuint32m2_t vamoswapei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m2_m))) +vuint32m2_t vamoswapei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m4))) +vuint32m4_t vamoswapei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m4_m))) +vuint32m4_t vamoswapei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m8))) +vuint32m8_t vamoswapei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m8_m))) +vuint32m8_t vamoswapei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32mf2))) +vuint32mf2_t vamoswapei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32mf2_m))) +vuint32mf2_t vamoswapei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32m1))) +vint32m1_t vamoswapei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32m1_m))) +vint32m1_t vamoswapei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32m2))) +vint32m2_t vamoswapei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32m2_m))) +vint32m2_t vamoswapei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32m4))) +vint32m4_t vamoswapei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32m4_m))) +vint32m4_t vamoswapei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32mf2))) +vint32mf2_t vamoswapei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32mf2_m))) +vint32mf2_t vamoswapei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32m1))) +vuint32m1_t vamoswapei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32m1_m))) +vuint32m1_t vamoswapei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32m2))) +vuint32m2_t vamoswapei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32m2_m))) +vuint32m2_t vamoswapei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32m4))) +vuint32m4_t vamoswapei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32m4_m))) +vuint32m4_t vamoswapei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32mf2))) +vuint32mf2_t vamoswapei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32mf2_m))) +vuint32mf2_t vamoswapei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m1))) +vint64m1_t vamoswapei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m1_m))) +vint64m1_t vamoswapei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m2))) +vint64m2_t vamoswapei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m2_m))) +vint64m2_t vamoswapei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m4))) +vint64m4_t vamoswapei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m4_m))) +vint64m4_t vamoswapei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m8))) +vint64m8_t vamoswapei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m8_m))) +vint64m8_t vamoswapei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m1))) +vuint64m1_t vamoswapei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m1_m))) +vuint64m1_t vamoswapei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m2))) +vuint64m2_t vamoswapei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m2_m))) +vuint64m2_t vamoswapei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m4))) +vuint64m4_t vamoswapei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m4_m))) +vuint64m4_t vamoswapei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m8))) +vuint64m8_t vamoswapei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m8_m))) +vuint64m8_t vamoswapei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m1))) +vint64m1_t vamoswapei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m1_m))) +vint64m1_t vamoswapei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m2))) +vint64m2_t vamoswapei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m2_m))) +vint64m2_t vamoswapei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m4))) +vint64m4_t vamoswapei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m4_m))) +vint64m4_t vamoswapei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m8))) +vint64m8_t vamoswapei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m8_m))) +vint64m8_t vamoswapei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m1))) +vuint64m1_t vamoswapei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m1_m))) +vuint64m1_t vamoswapei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m2))) +vuint64m2_t vamoswapei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m2_m))) +vuint64m2_t vamoswapei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m4))) +vuint64m4_t vamoswapei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m4_m))) +vuint64m4_t vamoswapei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m8))) +vuint64m8_t vamoswapei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m8_m))) +vuint64m8_t vamoswapei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m1))) +vint64m1_t vamoswapei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m1_m))) +vint64m1_t vamoswapei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m2))) +vint64m2_t vamoswapei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m2_m))) +vint64m2_t vamoswapei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m4))) +vint64m4_t vamoswapei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m4_m))) +vint64m4_t vamoswapei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m8))) +vint64m8_t vamoswapei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m8_m))) +vint64m8_t vamoswapei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m1))) +vuint64m1_t vamoswapei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m1_m))) +vuint64m1_t vamoswapei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m2))) +vuint64m2_t vamoswapei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m2_m))) +vuint64m2_t vamoswapei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m4))) +vuint64m4_t vamoswapei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m4_m))) +vuint64m4_t vamoswapei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m8))) +vuint64m8_t vamoswapei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m8_m))) +vuint64m8_t vamoswapei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m1))) +vint64m1_t vamoswapei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m1_m))) +vint64m1_t vamoswapei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m2))) +vint64m2_t vamoswapei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m2_m))) +vint64m2_t vamoswapei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m4))) +vint64m4_t vamoswapei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m4_m))) +vint64m4_t vamoswapei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m8))) +vint64m8_t vamoswapei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m8_m))) +vint64m8_t vamoswapei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m1))) +vuint64m1_t vamoswapei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m1_m))) +vuint64m1_t vamoswapei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m2))) +vuint64m2_t vamoswapei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m2_m))) +vuint64m2_t vamoswapei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m4))) +vuint64m4_t vamoswapei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m4_m))) +vuint64m4_t vamoswapei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m8))) +vuint64m8_t vamoswapei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m8_m))) +vuint64m8_t vamoswapei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m1))) +vint32m1_t vamoaddei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m1_m))) +vint32m1_t vamoaddei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m2))) +vint32m2_t vamoaddei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m2_m))) +vint32m2_t vamoaddei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m4))) +vint32m4_t vamoaddei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m4_m))) +vint32m4_t vamoaddei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m8))) +vint32m8_t vamoaddei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m8_m))) +vint32m8_t vamoaddei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32mf2))) +vint32mf2_t vamoaddei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32mf2_m))) +vint32mf2_t vamoaddei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m1))) +vuint32m1_t vamoaddei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m1_m))) +vuint32m1_t vamoaddei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m2))) +vuint32m2_t vamoaddei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m2_m))) +vuint32m2_t vamoaddei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m4))) +vuint32m4_t vamoaddei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m4_m))) +vuint32m4_t vamoaddei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m8))) +vuint32m8_t vamoaddei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m8_m))) +vuint32m8_t vamoaddei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32mf2))) +vuint32mf2_t vamoaddei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32mf2_m))) +vuint32mf2_t vamoaddei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m1))) +vint32m1_t vamoaddei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m1_m))) +vint32m1_t vamoaddei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m2))) +vint32m2_t vamoaddei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m2_m))) +vint32m2_t vamoaddei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m4))) +vint32m4_t vamoaddei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m4_m))) +vint32m4_t vamoaddei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m8))) +vint32m8_t vamoaddei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m8_m))) +vint32m8_t vamoaddei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32mf2))) +vint32mf2_t vamoaddei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32mf2_m))) +vint32mf2_t vamoaddei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m1))) +vuint32m1_t vamoaddei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m1_m))) +vuint32m1_t vamoaddei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m2))) +vuint32m2_t vamoaddei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m2_m))) +vuint32m2_t vamoaddei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m4))) +vuint32m4_t vamoaddei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m4_m))) +vuint32m4_t vamoaddei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m8))) +vuint32m8_t vamoaddei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m8_m))) +vuint32m8_t vamoaddei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32mf2))) +vuint32mf2_t vamoaddei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32mf2_m))) +vuint32mf2_t vamoaddei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m1))) +vint32m1_t vamoaddei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m1_m))) +vint32m1_t vamoaddei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m2))) +vint32m2_t vamoaddei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m2_m))) +vint32m2_t vamoaddei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m4))) +vint32m4_t vamoaddei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m4_m))) +vint32m4_t vamoaddei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m8))) +vint32m8_t vamoaddei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m8_m))) +vint32m8_t vamoaddei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32mf2))) +vint32mf2_t vamoaddei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32mf2_m))) +vint32mf2_t vamoaddei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m1))) +vuint32m1_t vamoaddei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m1_m))) +vuint32m1_t vamoaddei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m2))) +vuint32m2_t vamoaddei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m2_m))) +vuint32m2_t vamoaddei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m4))) +vuint32m4_t vamoaddei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m4_m))) +vuint32m4_t vamoaddei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m8))) +vuint32m8_t vamoaddei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m8_m))) +vuint32m8_t vamoaddei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32mf2))) +vuint32mf2_t vamoaddei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32mf2_m))) +vuint32mf2_t vamoaddei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32m1))) +vint32m1_t vamoaddei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32m1_m))) +vint32m1_t vamoaddei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32m2))) +vint32m2_t vamoaddei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32m2_m))) +vint32m2_t vamoaddei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32m4))) +vint32m4_t vamoaddei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32m4_m))) +vint32m4_t vamoaddei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32mf2))) +vint32mf2_t vamoaddei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32mf2_m))) +vint32mf2_t vamoaddei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32m1))) +vuint32m1_t vamoaddei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32m1_m))) +vuint32m1_t vamoaddei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32m2))) +vuint32m2_t vamoaddei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32m2_m))) +vuint32m2_t vamoaddei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32m4))) +vuint32m4_t vamoaddei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32m4_m))) +vuint32m4_t vamoaddei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32mf2))) +vuint32mf2_t vamoaddei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32mf2_m))) +vuint32mf2_t vamoaddei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m1))) +vint64m1_t vamoaddei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m1_m))) +vint64m1_t vamoaddei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m2))) +vint64m2_t vamoaddei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m2_m))) +vint64m2_t vamoaddei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m4))) +vint64m4_t vamoaddei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m4_m))) +vint64m4_t vamoaddei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m8))) +vint64m8_t vamoaddei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m8_m))) +vint64m8_t vamoaddei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m1))) +vuint64m1_t vamoaddei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m1_m))) +vuint64m1_t vamoaddei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m2))) +vuint64m2_t vamoaddei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m2_m))) +vuint64m2_t vamoaddei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m4))) +vuint64m4_t vamoaddei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m4_m))) +vuint64m4_t vamoaddei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m8))) +vuint64m8_t vamoaddei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m8_m))) +vuint64m8_t vamoaddei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m1))) +vint64m1_t vamoaddei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m1_m))) +vint64m1_t vamoaddei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m2))) +vint64m2_t vamoaddei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m2_m))) +vint64m2_t vamoaddei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m4))) +vint64m4_t vamoaddei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m4_m))) +vint64m4_t vamoaddei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m8))) +vint64m8_t vamoaddei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m8_m))) +vint64m8_t vamoaddei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m1))) +vuint64m1_t vamoaddei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m1_m))) +vuint64m1_t vamoaddei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m2))) +vuint64m2_t vamoaddei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m2_m))) +vuint64m2_t vamoaddei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m4))) +vuint64m4_t vamoaddei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m4_m))) +vuint64m4_t vamoaddei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m8))) +vuint64m8_t vamoaddei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m8_m))) +vuint64m8_t vamoaddei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m1))) +vint64m1_t vamoaddei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m1_m))) +vint64m1_t vamoaddei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m2))) +vint64m2_t vamoaddei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m2_m))) +vint64m2_t vamoaddei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m4))) +vint64m4_t vamoaddei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m4_m))) +vint64m4_t vamoaddei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m8))) +vint64m8_t vamoaddei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m8_m))) +vint64m8_t vamoaddei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m1))) +vuint64m1_t vamoaddei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m1_m))) +vuint64m1_t vamoaddei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m2))) +vuint64m2_t vamoaddei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m2_m))) +vuint64m2_t vamoaddei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m4))) +vuint64m4_t vamoaddei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m4_m))) +vuint64m4_t vamoaddei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m8))) +vuint64m8_t vamoaddei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m8_m))) +vuint64m8_t vamoaddei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m1))) +vint64m1_t vamoaddei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m1_m))) +vint64m1_t vamoaddei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m2))) +vint64m2_t vamoaddei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m2_m))) +vint64m2_t vamoaddei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m4))) +vint64m4_t vamoaddei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m4_m))) +vint64m4_t vamoaddei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m8))) +vint64m8_t vamoaddei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m8_m))) +vint64m8_t vamoaddei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m1))) +vuint64m1_t vamoaddei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m1_m))) +vuint64m1_t vamoaddei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m2))) +vuint64m2_t vamoaddei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m2_m))) +vuint64m2_t vamoaddei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m4))) +vuint64m4_t vamoaddei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m4_m))) +vuint64m4_t vamoaddei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m8))) +vuint64m8_t vamoaddei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m8_m))) +vuint64m8_t vamoaddei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m1))) +vint32m1_t vamoxorei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m1_m))) +vint32m1_t vamoxorei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m2))) +vint32m2_t vamoxorei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m2_m))) +vint32m2_t vamoxorei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m4))) +vint32m4_t vamoxorei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m4_m))) +vint32m4_t vamoxorei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m8))) +vint32m8_t vamoxorei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m8_m))) +vint32m8_t vamoxorei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32mf2))) +vint32mf2_t vamoxorei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32mf2_m))) +vint32mf2_t vamoxorei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m1))) +vuint32m1_t vamoxorei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m1_m))) +vuint32m1_t vamoxorei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m2))) +vuint32m2_t vamoxorei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m2_m))) +vuint32m2_t vamoxorei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m4))) +vuint32m4_t vamoxorei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m4_m))) +vuint32m4_t vamoxorei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m8))) +vuint32m8_t vamoxorei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m8_m))) +vuint32m8_t vamoxorei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32mf2))) +vuint32mf2_t vamoxorei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32mf2_m))) +vuint32mf2_t vamoxorei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m1))) +vint32m1_t vamoxorei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m1_m))) +vint32m1_t vamoxorei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m2))) +vint32m2_t vamoxorei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m2_m))) +vint32m2_t vamoxorei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m4))) +vint32m4_t vamoxorei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m4_m))) +vint32m4_t vamoxorei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m8))) +vint32m8_t vamoxorei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m8_m))) +vint32m8_t vamoxorei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32mf2))) +vint32mf2_t vamoxorei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32mf2_m))) +vint32mf2_t vamoxorei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m1))) +vuint32m1_t vamoxorei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m1_m))) +vuint32m1_t vamoxorei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m2))) +vuint32m2_t vamoxorei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m2_m))) +vuint32m2_t vamoxorei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m4))) +vuint32m4_t vamoxorei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m4_m))) +vuint32m4_t vamoxorei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m8))) +vuint32m8_t vamoxorei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m8_m))) +vuint32m8_t vamoxorei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32mf2))) +vuint32mf2_t vamoxorei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32mf2_m))) +vuint32mf2_t vamoxorei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m1))) +vint32m1_t vamoxorei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m1_m))) +vint32m1_t vamoxorei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m2))) +vint32m2_t vamoxorei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m2_m))) +vint32m2_t vamoxorei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m4))) +vint32m4_t vamoxorei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m4_m))) +vint32m4_t vamoxorei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m8))) +vint32m8_t vamoxorei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m8_m))) +vint32m8_t vamoxorei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32mf2))) +vint32mf2_t vamoxorei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32mf2_m))) +vint32mf2_t vamoxorei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m1))) +vuint32m1_t vamoxorei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m1_m))) +vuint32m1_t vamoxorei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m2))) +vuint32m2_t vamoxorei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m2_m))) +vuint32m2_t vamoxorei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m4))) +vuint32m4_t vamoxorei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m4_m))) +vuint32m4_t vamoxorei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m8))) +vuint32m8_t vamoxorei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m8_m))) +vuint32m8_t vamoxorei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32mf2))) +vuint32mf2_t vamoxorei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32mf2_m))) +vuint32mf2_t vamoxorei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32m1))) +vint32m1_t vamoxorei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32m1_m))) +vint32m1_t vamoxorei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32m2))) +vint32m2_t vamoxorei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32m2_m))) +vint32m2_t vamoxorei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32m4))) +vint32m4_t vamoxorei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32m4_m))) +vint32m4_t vamoxorei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32mf2))) +vint32mf2_t vamoxorei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32mf2_m))) +vint32mf2_t vamoxorei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32m1))) +vuint32m1_t vamoxorei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32m1_m))) +vuint32m1_t vamoxorei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32m2))) +vuint32m2_t vamoxorei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32m2_m))) +vuint32m2_t vamoxorei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32m4))) +vuint32m4_t vamoxorei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32m4_m))) +vuint32m4_t vamoxorei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32mf2))) +vuint32mf2_t vamoxorei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32mf2_m))) +vuint32mf2_t vamoxorei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m1))) +vint64m1_t vamoxorei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m1_m))) +vint64m1_t vamoxorei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m2))) +vint64m2_t vamoxorei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m2_m))) +vint64m2_t vamoxorei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m4))) +vint64m4_t vamoxorei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m4_m))) +vint64m4_t vamoxorei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m8))) +vint64m8_t vamoxorei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m8_m))) +vint64m8_t vamoxorei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m1))) +vuint64m1_t vamoxorei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m1_m))) +vuint64m1_t vamoxorei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m2))) +vuint64m2_t vamoxorei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m2_m))) +vuint64m2_t vamoxorei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m4))) +vuint64m4_t vamoxorei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m4_m))) +vuint64m4_t vamoxorei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m8))) +vuint64m8_t vamoxorei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m8_m))) +vuint64m8_t vamoxorei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m1))) +vint64m1_t vamoxorei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m1_m))) +vint64m1_t vamoxorei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m2))) +vint64m2_t vamoxorei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m2_m))) +vint64m2_t vamoxorei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m4))) +vint64m4_t vamoxorei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m4_m))) +vint64m4_t vamoxorei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m8))) +vint64m8_t vamoxorei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m8_m))) +vint64m8_t vamoxorei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m1))) +vuint64m1_t vamoxorei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m1_m))) +vuint64m1_t vamoxorei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m2))) +vuint64m2_t vamoxorei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m2_m))) +vuint64m2_t vamoxorei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m4))) +vuint64m4_t vamoxorei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m4_m))) +vuint64m4_t vamoxorei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m8))) +vuint64m8_t vamoxorei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m8_m))) +vuint64m8_t vamoxorei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m1))) +vint64m1_t vamoxorei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m1_m))) +vint64m1_t vamoxorei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m2))) +vint64m2_t vamoxorei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m2_m))) +vint64m2_t vamoxorei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m4))) +vint64m4_t vamoxorei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m4_m))) +vint64m4_t vamoxorei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m8))) +vint64m8_t vamoxorei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m8_m))) +vint64m8_t vamoxorei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m1))) +vuint64m1_t vamoxorei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m1_m))) +vuint64m1_t vamoxorei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m2))) +vuint64m2_t vamoxorei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m2_m))) +vuint64m2_t vamoxorei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m4))) +vuint64m4_t vamoxorei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m4_m))) +vuint64m4_t vamoxorei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m8))) +vuint64m8_t vamoxorei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m8_m))) +vuint64m8_t vamoxorei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m1))) +vint64m1_t vamoxorei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m1_m))) +vint64m1_t vamoxorei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m2))) +vint64m2_t vamoxorei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m2_m))) +vint64m2_t vamoxorei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m4))) +vint64m4_t vamoxorei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m4_m))) +vint64m4_t vamoxorei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m8))) +vint64m8_t vamoxorei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m8_m))) +vint64m8_t vamoxorei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m1))) +vuint64m1_t vamoxorei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m1_m))) +vuint64m1_t vamoxorei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m2))) +vuint64m2_t vamoxorei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m2_m))) +vuint64m2_t vamoxorei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m4))) +vuint64m4_t vamoxorei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m4_m))) +vuint64m4_t vamoxorei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m8))) +vuint64m8_t vamoxorei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m8_m))) +vuint64m8_t vamoxorei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m1))) +vint32m1_t vamoandei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m1_m))) +vint32m1_t vamoandei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m2))) +vint32m2_t vamoandei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m2_m))) +vint32m2_t vamoandei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m4))) +vint32m4_t vamoandei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m4_m))) +vint32m4_t vamoandei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m8))) +vint32m8_t vamoandei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m8_m))) +vint32m8_t vamoandei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32mf2))) +vint32mf2_t vamoandei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32mf2_m))) +vint32mf2_t vamoandei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m1))) +vuint32m1_t vamoandei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m1_m))) +vuint32m1_t vamoandei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m2))) +vuint32m2_t vamoandei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m2_m))) +vuint32m2_t vamoandei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m4))) +vuint32m4_t vamoandei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m4_m))) +vuint32m4_t vamoandei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m8))) +vuint32m8_t vamoandei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m8_m))) +vuint32m8_t vamoandei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32mf2))) +vuint32mf2_t vamoandei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32mf2_m))) +vuint32mf2_t vamoandei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m1))) +vint32m1_t vamoandei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m1_m))) +vint32m1_t vamoandei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m2))) +vint32m2_t vamoandei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m2_m))) +vint32m2_t vamoandei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m4))) +vint32m4_t vamoandei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m4_m))) +vint32m4_t vamoandei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m8))) +vint32m8_t vamoandei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m8_m))) +vint32m8_t vamoandei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32mf2))) +vint32mf2_t vamoandei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32mf2_m))) +vint32mf2_t vamoandei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m1))) +vuint32m1_t vamoandei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m1_m))) +vuint32m1_t vamoandei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m2))) +vuint32m2_t vamoandei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m2_m))) +vuint32m2_t vamoandei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m4))) +vuint32m4_t vamoandei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m4_m))) +vuint32m4_t vamoandei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m8))) +vuint32m8_t vamoandei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m8_m))) +vuint32m8_t vamoandei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32mf2))) +vuint32mf2_t vamoandei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32mf2_m))) +vuint32mf2_t vamoandei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m1))) +vint32m1_t vamoandei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m1_m))) +vint32m1_t vamoandei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m2))) +vint32m2_t vamoandei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m2_m))) +vint32m2_t vamoandei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m4))) +vint32m4_t vamoandei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m4_m))) +vint32m4_t vamoandei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m8))) +vint32m8_t vamoandei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m8_m))) +vint32m8_t vamoandei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32mf2))) +vint32mf2_t vamoandei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32mf2_m))) +vint32mf2_t vamoandei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m1))) +vuint32m1_t vamoandei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m1_m))) +vuint32m1_t vamoandei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m2))) +vuint32m2_t vamoandei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m2_m))) +vuint32m2_t vamoandei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m4))) +vuint32m4_t vamoandei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m4_m))) +vuint32m4_t vamoandei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m8))) +vuint32m8_t vamoandei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m8_m))) +vuint32m8_t vamoandei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32mf2))) +vuint32mf2_t vamoandei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32mf2_m))) +vuint32mf2_t vamoandei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32m1))) +vint32m1_t vamoandei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32m1_m))) +vint32m1_t vamoandei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32m2))) +vint32m2_t vamoandei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32m2_m))) +vint32m2_t vamoandei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32m4))) +vint32m4_t vamoandei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32m4_m))) +vint32m4_t vamoandei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32mf2))) +vint32mf2_t vamoandei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32mf2_m))) +vint32mf2_t vamoandei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32m1))) +vuint32m1_t vamoandei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32m1_m))) +vuint32m1_t vamoandei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32m2))) +vuint32m2_t vamoandei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32m2_m))) +vuint32m2_t vamoandei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32m4))) +vuint32m4_t vamoandei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32m4_m))) +vuint32m4_t vamoandei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32mf2))) +vuint32mf2_t vamoandei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32mf2_m))) +vuint32mf2_t vamoandei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m1))) +vint64m1_t vamoandei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m1_m))) +vint64m1_t vamoandei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m2))) +vint64m2_t vamoandei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m2_m))) +vint64m2_t vamoandei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m4))) +vint64m4_t vamoandei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m4_m))) +vint64m4_t vamoandei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m8))) +vint64m8_t vamoandei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m8_m))) +vint64m8_t vamoandei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m1))) +vuint64m1_t vamoandei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m1_m))) +vuint64m1_t vamoandei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m2))) +vuint64m2_t vamoandei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m2_m))) +vuint64m2_t vamoandei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m4))) +vuint64m4_t vamoandei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m4_m))) +vuint64m4_t vamoandei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m8))) +vuint64m8_t vamoandei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m8_m))) +vuint64m8_t vamoandei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m1))) +vint64m1_t vamoandei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m1_m))) +vint64m1_t vamoandei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m2))) +vint64m2_t vamoandei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m2_m))) +vint64m2_t vamoandei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m4))) +vint64m4_t vamoandei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m4_m))) +vint64m4_t vamoandei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m8))) +vint64m8_t vamoandei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m8_m))) +vint64m8_t vamoandei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m1))) +vuint64m1_t vamoandei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m1_m))) +vuint64m1_t vamoandei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m2))) +vuint64m2_t vamoandei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m2_m))) +vuint64m2_t vamoandei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m4))) +vuint64m4_t vamoandei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m4_m))) +vuint64m4_t vamoandei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m8))) +vuint64m8_t vamoandei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m8_m))) +vuint64m8_t vamoandei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m1))) +vint64m1_t vamoandei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m1_m))) +vint64m1_t vamoandei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m2))) +vint64m2_t vamoandei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m2_m))) +vint64m2_t vamoandei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m4))) +vint64m4_t vamoandei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m4_m))) +vint64m4_t vamoandei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m8))) +vint64m8_t vamoandei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m8_m))) +vint64m8_t vamoandei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m1))) +vuint64m1_t vamoandei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m1_m))) +vuint64m1_t vamoandei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m2))) +vuint64m2_t vamoandei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m2_m))) +vuint64m2_t vamoandei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m4))) +vuint64m4_t vamoandei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m4_m))) +vuint64m4_t vamoandei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m8))) +vuint64m8_t vamoandei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m8_m))) +vuint64m8_t vamoandei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m1))) +vint64m1_t vamoandei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m1_m))) +vint64m1_t vamoandei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m2))) +vint64m2_t vamoandei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m2_m))) +vint64m2_t vamoandei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m4))) +vint64m4_t vamoandei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m4_m))) +vint64m4_t vamoandei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m8))) +vint64m8_t vamoandei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m8_m))) +vint64m8_t vamoandei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m1))) +vuint64m1_t vamoandei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m1_m))) +vuint64m1_t vamoandei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m2))) +vuint64m2_t vamoandei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m2_m))) +vuint64m2_t vamoandei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m4))) +vuint64m4_t vamoandei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m4_m))) +vuint64m4_t vamoandei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m8))) +vuint64m8_t vamoandei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m8_m))) +vuint64m8_t vamoandei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m1))) +vint32m1_t vamoorei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m1_m))) +vint32m1_t vamoorei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m2))) +vint32m2_t vamoorei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m2_m))) +vint32m2_t vamoorei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m4))) +vint32m4_t vamoorei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m4_m))) +vint32m4_t vamoorei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m8))) +vint32m8_t vamoorei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m8_m))) +vint32m8_t vamoorei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32mf2))) +vint32mf2_t vamoorei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32mf2_m))) +vint32mf2_t vamoorei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m1))) +vuint32m1_t vamoorei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m1_m))) +vuint32m1_t vamoorei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m2))) +vuint32m2_t vamoorei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m2_m))) +vuint32m2_t vamoorei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m4))) +vuint32m4_t vamoorei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m4_m))) +vuint32m4_t vamoorei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m8))) +vuint32m8_t vamoorei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m8_m))) +vuint32m8_t vamoorei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32mf2))) +vuint32mf2_t vamoorei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32mf2_m))) +vuint32mf2_t vamoorei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m1))) +vint32m1_t vamoorei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m1_m))) +vint32m1_t vamoorei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m2))) +vint32m2_t vamoorei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m2_m))) +vint32m2_t vamoorei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m4))) +vint32m4_t vamoorei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m4_m))) +vint32m4_t vamoorei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m8))) +vint32m8_t vamoorei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m8_m))) +vint32m8_t vamoorei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32mf2))) +vint32mf2_t vamoorei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32mf2_m))) +vint32mf2_t vamoorei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m1))) +vuint32m1_t vamoorei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m1_m))) +vuint32m1_t vamoorei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m2))) +vuint32m2_t vamoorei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m2_m))) +vuint32m2_t vamoorei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m4))) +vuint32m4_t vamoorei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m4_m))) +vuint32m4_t vamoorei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m8))) +vuint32m8_t vamoorei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m8_m))) +vuint32m8_t vamoorei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32mf2))) +vuint32mf2_t vamoorei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32mf2_m))) +vuint32mf2_t vamoorei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m1))) +vint32m1_t vamoorei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m1_m))) +vint32m1_t vamoorei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m2))) +vint32m2_t vamoorei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m2_m))) +vint32m2_t vamoorei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m4))) +vint32m4_t vamoorei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m4_m))) +vint32m4_t vamoorei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m8))) +vint32m8_t vamoorei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m8_m))) +vint32m8_t vamoorei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32mf2))) +vint32mf2_t vamoorei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32mf2_m))) +vint32mf2_t vamoorei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m1))) +vuint32m1_t vamoorei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m1_m))) +vuint32m1_t vamoorei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m2))) +vuint32m2_t vamoorei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m2_m))) +vuint32m2_t vamoorei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m4))) +vuint32m4_t vamoorei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m4_m))) +vuint32m4_t vamoorei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m8))) +vuint32m8_t vamoorei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m8_m))) +vuint32m8_t vamoorei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32mf2))) +vuint32mf2_t vamoorei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32mf2_m))) +vuint32mf2_t vamoorei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32m1))) +vint32m1_t vamoorei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32m1_m))) +vint32m1_t vamoorei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32m2))) +vint32m2_t vamoorei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32m2_m))) +vint32m2_t vamoorei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32m4))) +vint32m4_t vamoorei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32m4_m))) +vint32m4_t vamoorei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32mf2))) +vint32mf2_t vamoorei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32mf2_m))) +vint32mf2_t vamoorei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32m1))) +vuint32m1_t vamoorei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32m1_m))) +vuint32m1_t vamoorei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32m2))) +vuint32m2_t vamoorei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32m2_m))) +vuint32m2_t vamoorei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32m4))) +vuint32m4_t vamoorei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32m4_m))) +vuint32m4_t vamoorei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32mf2))) +vuint32mf2_t vamoorei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32mf2_m))) +vuint32mf2_t vamoorei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m1))) +vint64m1_t vamoorei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m1_m))) +vint64m1_t vamoorei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m2))) +vint64m2_t vamoorei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m2_m))) +vint64m2_t vamoorei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m4))) +vint64m4_t vamoorei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m4_m))) +vint64m4_t vamoorei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m8))) +vint64m8_t vamoorei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m8_m))) +vint64m8_t vamoorei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m1))) +vuint64m1_t vamoorei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m1_m))) +vuint64m1_t vamoorei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m2))) +vuint64m2_t vamoorei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m2_m))) +vuint64m2_t vamoorei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m4))) +vuint64m4_t vamoorei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m4_m))) +vuint64m4_t vamoorei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m8))) +vuint64m8_t vamoorei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m8_m))) +vuint64m8_t vamoorei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m1))) +vint64m1_t vamoorei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m1_m))) +vint64m1_t vamoorei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m2))) +vint64m2_t vamoorei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m2_m))) +vint64m2_t vamoorei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m4))) +vint64m4_t vamoorei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m4_m))) +vint64m4_t vamoorei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m8))) +vint64m8_t vamoorei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m8_m))) +vint64m8_t vamoorei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m1))) +vuint64m1_t vamoorei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m1_m))) +vuint64m1_t vamoorei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m2))) +vuint64m2_t vamoorei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m2_m))) +vuint64m2_t vamoorei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m4))) +vuint64m4_t vamoorei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m4_m))) +vuint64m4_t vamoorei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m8))) +vuint64m8_t vamoorei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m8_m))) +vuint64m8_t vamoorei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m1))) +vint64m1_t vamoorei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m1_m))) +vint64m1_t vamoorei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m2))) +vint64m2_t vamoorei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m2_m))) +vint64m2_t vamoorei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m4))) +vint64m4_t vamoorei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m4_m))) +vint64m4_t vamoorei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m8))) +vint64m8_t vamoorei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m8_m))) +vint64m8_t vamoorei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m1))) +vuint64m1_t vamoorei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m1_m))) +vuint64m1_t vamoorei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m2))) +vuint64m2_t vamoorei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m2_m))) +vuint64m2_t vamoorei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m4))) +vuint64m4_t vamoorei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m4_m))) +vuint64m4_t vamoorei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m8))) +vuint64m8_t vamoorei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m8_m))) +vuint64m8_t vamoorei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m1))) +vint64m1_t vamoorei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m1_m))) +vint64m1_t vamoorei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m2))) +vint64m2_t vamoorei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m2_m))) +vint64m2_t vamoorei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m4))) +vint64m4_t vamoorei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m4_m))) +vint64m4_t vamoorei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m8))) +vint64m8_t vamoorei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m8_m))) +vint64m8_t vamoorei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m1))) +vuint64m1_t vamoorei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m1_m))) +vuint64m1_t vamoorei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m2))) +vuint64m2_t vamoorei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m2_m))) +vuint64m2_t vamoorei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m4))) +vuint64m4_t vamoorei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m4_m))) +vuint64m4_t vamoorei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m8))) +vuint64m8_t vamoorei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m8_m))) +vuint64m8_t vamoorei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m1))) +vint32m1_t vamominei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m1_m))) +vint32m1_t vamominei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m2))) +vint32m2_t vamominei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m2_m))) +vint32m2_t vamominei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m4))) +vint32m4_t vamominei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m4_m))) +vint32m4_t vamominei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m8))) +vint32m8_t vamominei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m8_m))) +vint32m8_t vamominei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32mf2))) +vint32mf2_t vamominei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32mf2_m))) +vint32mf2_t vamominei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m1))) +vint32m1_t vamominei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m1_m))) +vint32m1_t vamominei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m2))) +vint32m2_t vamominei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m2_m))) +vint32m2_t vamominei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m4))) +vint32m4_t vamominei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m4_m))) +vint32m4_t vamominei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m8))) +vint32m8_t vamominei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m8_m))) +vint32m8_t vamominei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32mf2))) +vint32mf2_t vamominei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32mf2_m))) +vint32mf2_t vamominei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m1))) +vint32m1_t vamominei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m1_m))) +vint32m1_t vamominei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m2))) +vint32m2_t vamominei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m2_m))) +vint32m2_t vamominei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m4))) +vint32m4_t vamominei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m4_m))) +vint32m4_t vamominei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m8))) +vint32m8_t vamominei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m8_m))) +vint32m8_t vamominei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32mf2))) +vint32mf2_t vamominei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32mf2_m))) +vint32mf2_t vamominei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32m1))) +vint32m1_t vamominei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32m1_m))) +vint32m1_t vamominei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32m2))) +vint32m2_t vamominei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32m2_m))) +vint32m2_t vamominei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32m4))) +vint32m4_t vamominei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32m4_m))) +vint32m4_t vamominei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32mf2))) +vint32mf2_t vamominei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32mf2_m))) +vint32mf2_t vamominei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m1))) +vint64m1_t vamominei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m1_m))) +vint64m1_t vamominei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m2))) +vint64m2_t vamominei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m2_m))) +vint64m2_t vamominei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m4))) +vint64m4_t vamominei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m4_m))) +vint64m4_t vamominei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m8))) +vint64m8_t vamominei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m8_m))) +vint64m8_t vamominei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m1))) +vint64m1_t vamominei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m1_m))) +vint64m1_t vamominei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m2))) +vint64m2_t vamominei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m2_m))) +vint64m2_t vamominei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m4))) +vint64m4_t vamominei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m4_m))) +vint64m4_t vamominei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m8))) +vint64m8_t vamominei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m8_m))) +vint64m8_t vamominei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m1))) +vint64m1_t vamominei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m1_m))) +vint64m1_t vamominei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m2))) +vint64m2_t vamominei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m2_m))) +vint64m2_t vamominei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m4))) +vint64m4_t vamominei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m4_m))) +vint64m4_t vamominei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m8))) +vint64m8_t vamominei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m8_m))) +vint64m8_t vamominei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m1))) +vint64m1_t vamominei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m1_m))) +vint64m1_t vamominei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m2))) +vint64m2_t vamominei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m2_m))) +vint64m2_t vamominei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m4))) +vint64m4_t vamominei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m4_m))) +vint64m4_t vamominei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m8))) +vint64m8_t vamominei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m8_m))) +vint64m8_t vamominei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m1))) +vint32m1_t vamomaxei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m1_m))) +vint32m1_t vamomaxei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m2))) +vint32m2_t vamomaxei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m2_m))) +vint32m2_t vamomaxei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m4))) +vint32m4_t vamomaxei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m4_m))) +vint32m4_t vamomaxei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m8))) +vint32m8_t vamomaxei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m8_m))) +vint32m8_t vamomaxei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32mf2))) +vint32mf2_t vamomaxei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32mf2_m))) +vint32mf2_t vamomaxei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m1))) +vint32m1_t vamomaxei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m1_m))) +vint32m1_t vamomaxei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m2))) +vint32m2_t vamomaxei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m2_m))) +vint32m2_t vamomaxei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m4))) +vint32m4_t vamomaxei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m4_m))) +vint32m4_t vamomaxei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m8))) +vint32m8_t vamomaxei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m8_m))) +vint32m8_t vamomaxei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32mf2))) +vint32mf2_t vamomaxei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32mf2_m))) +vint32mf2_t vamomaxei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m1))) +vint32m1_t vamomaxei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m1_m))) +vint32m1_t vamomaxei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m2))) +vint32m2_t vamomaxei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m2_m))) +vint32m2_t vamomaxei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m4))) +vint32m4_t vamomaxei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m4_m))) +vint32m4_t vamomaxei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m8))) +vint32m8_t vamomaxei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m8_m))) +vint32m8_t vamomaxei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32mf2))) +vint32mf2_t vamomaxei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32mf2_m))) +vint32mf2_t vamomaxei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32m1))) +vint32m1_t vamomaxei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32m1_m))) +vint32m1_t vamomaxei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32m2))) +vint32m2_t vamomaxei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32m2_m))) +vint32m2_t vamomaxei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32m4))) +vint32m4_t vamomaxei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32m4_m))) +vint32m4_t vamomaxei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32mf2))) +vint32mf2_t vamomaxei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32mf2_m))) +vint32mf2_t vamomaxei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m1))) +vint64m1_t vamomaxei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m1_m))) +vint64m1_t vamomaxei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m2))) +vint64m2_t vamomaxei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m2_m))) +vint64m2_t vamomaxei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m4))) +vint64m4_t vamomaxei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m4_m))) +vint64m4_t vamomaxei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m8))) +vint64m8_t vamomaxei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m8_m))) +vint64m8_t vamomaxei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m1))) +vint64m1_t vamomaxei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m1_m))) +vint64m1_t vamomaxei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m2))) +vint64m2_t vamomaxei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m2_m))) +vint64m2_t vamomaxei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m4))) +vint64m4_t vamomaxei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m4_m))) +vint64m4_t vamomaxei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m8))) +vint64m8_t vamomaxei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m8_m))) +vint64m8_t vamomaxei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m1))) +vint64m1_t vamomaxei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m1_m))) +vint64m1_t vamomaxei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m2))) +vint64m2_t vamomaxei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m2_m))) +vint64m2_t vamomaxei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m4))) +vint64m4_t vamomaxei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m4_m))) +vint64m4_t vamomaxei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m8))) +vint64m8_t vamomaxei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m8_m))) +vint64m8_t vamomaxei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m1))) +vint64m1_t vamomaxei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m1_m))) +vint64m1_t vamomaxei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m2))) +vint64m2_t vamomaxei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m2_m))) +vint64m2_t vamomaxei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m4))) +vint64m4_t vamomaxei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m4_m))) +vint64m4_t vamomaxei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m8))) +vint64m8_t vamomaxei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m8_m))) +vint64m8_t vamomaxei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m1))) +vuint32m1_t vamominuei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m1_m))) +vuint32m1_t vamominuei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m2))) +vuint32m2_t vamominuei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m2_m))) +vuint32m2_t vamominuei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m4))) +vuint32m4_t vamominuei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m4_m))) +vuint32m4_t vamominuei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m8))) +vuint32m8_t vamominuei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m8_m))) +vuint32m8_t vamominuei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32mf2))) +vuint32mf2_t vamominuei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32mf2_m))) +vuint32mf2_t vamominuei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m1))) +vuint32m1_t vamominuei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m1_m))) +vuint32m1_t vamominuei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m2))) +vuint32m2_t vamominuei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m2_m))) +vuint32m2_t vamominuei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m4))) +vuint32m4_t vamominuei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m4_m))) +vuint32m4_t vamominuei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m8))) +vuint32m8_t vamominuei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m8_m))) +vuint32m8_t vamominuei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32mf2))) +vuint32mf2_t vamominuei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32mf2_m))) +vuint32mf2_t vamominuei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m1))) +vuint32m1_t vamominuei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m1_m))) +vuint32m1_t vamominuei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m2))) +vuint32m2_t vamominuei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m2_m))) +vuint32m2_t vamominuei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m4))) +vuint32m4_t vamominuei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m4_m))) +vuint32m4_t vamominuei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m8))) +vuint32m8_t vamominuei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m8_m))) +vuint32m8_t vamominuei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32mf2))) +vuint32mf2_t vamominuei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32mf2_m))) +vuint32mf2_t vamominuei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32m1))) +vuint32m1_t vamominuei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32m1_m))) +vuint32m1_t vamominuei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32m2))) +vuint32m2_t vamominuei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32m2_m))) +vuint32m2_t vamominuei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32m4))) +vuint32m4_t vamominuei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32m4_m))) +vuint32m4_t vamominuei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32mf2))) +vuint32mf2_t vamominuei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32mf2_m))) +vuint32mf2_t vamominuei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m1))) +vuint64m1_t vamominuei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m1_m))) +vuint64m1_t vamominuei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m2))) +vuint64m2_t vamominuei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m2_m))) +vuint64m2_t vamominuei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m4))) +vuint64m4_t vamominuei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m4_m))) +vuint64m4_t vamominuei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m8))) +vuint64m8_t vamominuei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m8_m))) +vuint64m8_t vamominuei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m1))) +vuint64m1_t vamominuei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m1_m))) +vuint64m1_t vamominuei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m2))) +vuint64m2_t vamominuei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m2_m))) +vuint64m2_t vamominuei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m4))) +vuint64m4_t vamominuei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m4_m))) +vuint64m4_t vamominuei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m8))) +vuint64m8_t vamominuei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m8_m))) +vuint64m8_t vamominuei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m1))) +vuint64m1_t vamominuei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m1_m))) +vuint64m1_t vamominuei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m2))) +vuint64m2_t vamominuei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m2_m))) +vuint64m2_t vamominuei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m4))) +vuint64m4_t vamominuei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m4_m))) +vuint64m4_t vamominuei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m8))) +vuint64m8_t vamominuei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m8_m))) +vuint64m8_t vamominuei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m1))) +vuint64m1_t vamominuei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m1_m))) +vuint64m1_t vamominuei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m2))) +vuint64m2_t vamominuei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m2_m))) +vuint64m2_t vamominuei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m4))) +vuint64m4_t vamominuei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m4_m))) +vuint64m4_t vamominuei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m8))) +vuint64m8_t vamominuei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m8_m))) +vuint64m8_t vamominuei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m1))) +vuint32m1_t vamomaxuei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m1_m))) +vuint32m1_t vamomaxuei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m2))) +vuint32m2_t vamomaxuei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m2_m))) +vuint32m2_t vamomaxuei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m4))) +vuint32m4_t vamomaxuei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m4_m))) +vuint32m4_t vamomaxuei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m8))) +vuint32m8_t vamomaxuei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m8_m))) +vuint32m8_t vamomaxuei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32mf2))) +vuint32mf2_t vamomaxuei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32mf2_m))) +vuint32mf2_t vamomaxuei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m1))) +vuint32m1_t vamomaxuei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m1_m))) +vuint32m1_t vamomaxuei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m2))) +vuint32m2_t vamomaxuei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m2_m))) +vuint32m2_t vamomaxuei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m4))) +vuint32m4_t vamomaxuei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m4_m))) +vuint32m4_t vamomaxuei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m8))) +vuint32m8_t vamomaxuei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m8_m))) +vuint32m8_t vamomaxuei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32mf2))) +vuint32mf2_t vamomaxuei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32mf2_m))) +vuint32mf2_t vamomaxuei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m1))) +vuint32m1_t vamomaxuei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m1_m))) +vuint32m1_t vamomaxuei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m2))) +vuint32m2_t vamomaxuei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m2_m))) +vuint32m2_t vamomaxuei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m4))) +vuint32m4_t vamomaxuei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m4_m))) +vuint32m4_t vamomaxuei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m8))) +vuint32m8_t vamomaxuei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m8_m))) +vuint32m8_t vamomaxuei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32mf2))) +vuint32mf2_t vamomaxuei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32mf2_m))) +vuint32mf2_t vamomaxuei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32m1))) +vuint32m1_t vamomaxuei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32m1_m))) +vuint32m1_t vamomaxuei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32m2))) +vuint32m2_t vamomaxuei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32m2_m))) +vuint32m2_t vamomaxuei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32m4))) +vuint32m4_t vamomaxuei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32m4_m))) +vuint32m4_t vamomaxuei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32mf2))) +vuint32mf2_t vamomaxuei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32mf2_m))) +vuint32mf2_t vamomaxuei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m1))) +vuint64m1_t vamomaxuei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m1_m))) +vuint64m1_t vamomaxuei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m2))) +vuint64m2_t vamomaxuei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m2_m))) +vuint64m2_t vamomaxuei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m4))) +vuint64m4_t vamomaxuei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m4_m))) +vuint64m4_t vamomaxuei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m8))) +vuint64m8_t vamomaxuei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m8_m))) +vuint64m8_t vamomaxuei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m1))) +vuint64m1_t vamomaxuei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m1_m))) +vuint64m1_t vamomaxuei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m2))) +vuint64m2_t vamomaxuei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m2_m))) +vuint64m2_t vamomaxuei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m4))) +vuint64m4_t vamomaxuei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m4_m))) +vuint64m4_t vamomaxuei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m8))) +vuint64m8_t vamomaxuei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m8_m))) +vuint64m8_t vamomaxuei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m1))) +vuint64m1_t vamomaxuei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m1_m))) +vuint64m1_t vamomaxuei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m2))) +vuint64m2_t vamomaxuei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m2_m))) +vuint64m2_t vamomaxuei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m4))) +vuint64m4_t vamomaxuei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m4_m))) +vuint64m4_t vamomaxuei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m8))) +vuint64m8_t vamomaxuei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m8_m))) +vuint64m8_t vamomaxuei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m1))) +vuint64m1_t vamomaxuei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m1_m))) +vuint64m1_t vamomaxuei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m2))) +vuint64m2_t vamomaxuei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m2_m))) +vuint64m2_t vamomaxuei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m4))) +vuint64m4_t vamomaxuei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m4_m))) +vuint64m4_t vamomaxuei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m8))) +vuint64m8_t vamomaxuei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m8_m))) +vuint64m8_t vamomaxuei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m1))) +vint32m1_t vamoswapei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m1_m))) +vint32m1_t vamoswapei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m2))) +vint32m2_t vamoswapei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m2_m))) +vint32m2_t vamoswapei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m4))) +vint32m4_t vamoswapei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m4_m))) +vint32m4_t vamoswapei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m8))) +vint32m8_t vamoswapei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m8_m))) +vint32m8_t vamoswapei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32mf2))) +vint32mf2_t vamoswapei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32mf2_m))) +vint32mf2_t vamoswapei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m1))) +vuint32m1_t vamoswapei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m1_m))) +vuint32m1_t vamoswapei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m2))) +vuint32m2_t vamoswapei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m2_m))) +vuint32m2_t vamoswapei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m4))) +vuint32m4_t vamoswapei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m4_m))) +vuint32m4_t vamoswapei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m8))) +vuint32m8_t vamoswapei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m8_m))) +vuint32m8_t vamoswapei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32mf2))) +vuint32mf2_t vamoswapei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32mf2_m))) +vuint32mf2_t vamoswapei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +#endif + +#if defined(__riscv_f) && defined(__riscv_zvamo) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m1))) +vfloat32m1_t vamoswapei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m1_m))) +vfloat32m1_t vamoswapei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m2))) +vfloat32m2_t vamoswapei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m2_m))) +vfloat32m2_t vamoswapei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m4))) +vfloat32m4_t vamoswapei8(float * op0, vuint8m1_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m4_m))) +vfloat32m4_t vamoswapei8(vbool8_t op0, float * op1, vuint8m1_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m8))) +vfloat32m8_t vamoswapei8(float * op0, vuint8m2_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m8_m))) +vfloat32m8_t vamoswapei8(vbool4_t op0, float * op1, vuint8m2_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32mf2))) +vfloat32mf2_t vamoswapei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32mf2_m))) +vfloat32mf2_t vamoswapei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m1))) +vfloat32m1_t vamoswapei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m1_m))) +vfloat32m1_t vamoswapei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m2))) +vfloat32m2_t vamoswapei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m2_m))) +vfloat32m2_t vamoswapei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m4))) +vfloat32m4_t vamoswapei16(float * op0, vuint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m4_m))) +vfloat32m4_t vamoswapei16(vbool8_t op0, float * op1, vuint16m2_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m8))) +vfloat32m8_t vamoswapei16(float * op0, vuint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m8_m))) +vfloat32m8_t vamoswapei16(vbool4_t op0, float * op1, vuint16m4_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32mf2))) +vfloat32mf2_t vamoswapei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32mf2_m))) +vfloat32mf2_t vamoswapei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m1))) +vfloat32m1_t vamoswapei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m1_m))) +vfloat32m1_t vamoswapei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m2))) +vfloat32m2_t vamoswapei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m2_m))) +vfloat32m2_t vamoswapei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m4))) +vfloat32m4_t vamoswapei32(float * op0, vuint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m4_m))) +vfloat32m4_t vamoswapei32(vbool8_t op0, float * op1, vuint32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m8))) +vfloat32m8_t vamoswapei32(float * op0, vuint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m8_m))) +vfloat32m8_t vamoswapei32(vbool4_t op0, float * op1, vuint32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32mf2))) +vfloat32mf2_t vamoswapei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32mf2_m))) +vfloat32mf2_t vamoswapei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32m1))) +vfloat32m1_t vamoswapei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32m1_m))) +vfloat32m1_t vamoswapei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32m2))) +vfloat32m2_t vamoswapei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32m2_m))) +vfloat32m2_t vamoswapei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32m4))) +vfloat32m4_t vamoswapei64(float * op0, vuint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32m4_m))) +vfloat32m4_t vamoswapei64(vbool8_t op0, float * op1, vuint64m8_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32mf2))) +vfloat32mf2_t vamoswapei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32mf2_m))) +vfloat32mf2_t vamoswapei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, size_t op4); + +#endif + +#if defined(__riscv_d) && defined(__riscv_zvamo) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m1))) +vfloat64m1_t vamoswapei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m1_m))) +vfloat64m1_t vamoswapei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m2))) +vfloat64m2_t vamoswapei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m2_m))) +vfloat64m2_t vamoswapei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m4))) +vfloat64m4_t vamoswapei8(double * op0, vuint8mf2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m4_m))) +vfloat64m4_t vamoswapei8(vbool16_t op0, double * op1, vuint8mf2_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m8))) +vfloat64m8_t vamoswapei8(double * op0, vuint8m1_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m8_m))) +vfloat64m8_t vamoswapei8(vbool8_t op0, double * op1, vuint8m1_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m1))) +vfloat64m1_t vamoswapei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m1_m))) +vfloat64m1_t vamoswapei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m2))) +vfloat64m2_t vamoswapei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m2_m))) +vfloat64m2_t vamoswapei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m4))) +vfloat64m4_t vamoswapei16(double * op0, vuint16m1_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m4_m))) +vfloat64m4_t vamoswapei16(vbool16_t op0, double * op1, vuint16m1_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m8))) +vfloat64m8_t vamoswapei16(double * op0, vuint16m2_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m8_m))) +vfloat64m8_t vamoswapei16(vbool8_t op0, double * op1, vuint16m2_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m1))) +vfloat64m1_t vamoswapei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m1_m))) +vfloat64m1_t vamoswapei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m2))) +vfloat64m2_t vamoswapei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m2_m))) +vfloat64m2_t vamoswapei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m4))) +vfloat64m4_t vamoswapei32(double * op0, vuint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m4_m))) +vfloat64m4_t vamoswapei32(vbool16_t op0, double * op1, vuint32m2_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m8))) +vfloat64m8_t vamoswapei32(double * op0, vuint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m8_m))) +vfloat64m8_t vamoswapei32(vbool8_t op0, double * op1, vuint32m4_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m1))) +vfloat64m1_t vamoswapei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m1_m))) +vfloat64m1_t vamoswapei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m2))) +vfloat64m2_t vamoswapei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m2_m))) +vfloat64m2_t vamoswapei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m4))) +vfloat64m4_t vamoswapei64(double * op0, vuint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m4_m))) +vfloat64m4_t vamoswapei64(vbool16_t op0, double * op1, vuint64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m8))) +vfloat64m8_t vamoswapei64(double * op0, vuint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m8_m))) +vfloat64m8_t vamoswapei64(vbool8_t op0, double * op1, vuint64m8_t op2, vfloat64m8_t op3, size_t op4); + +#endif + +#if defined(__riscv_zvlsseg) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i64m1))) +void vloxseg2ei8(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i64m1_m))) +void vloxseg2ei8(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i64m2))) +void vloxseg2ei8(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i64m2_m))) +void vloxseg2ei8(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i64m4))) +void vloxseg2ei8(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i64m4_m))) +void vloxseg2ei8(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u64m1))) +void vloxseg2ei8(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u64m1_m))) +void vloxseg2ei8(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u64m2))) +void vloxseg2ei8(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u64m2_m))) +void vloxseg2ei8(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u64m4))) +void vloxseg2ei8(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u64m4_m))) +void vloxseg2ei8(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i64m1))) +void vloxseg3ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i64m1_m))) +void vloxseg3ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i64m2))) +void vloxseg3ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i64m2_m))) +void vloxseg3ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u64m1))) +void vloxseg3ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u64m1_m))) +void vloxseg3ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u64m2))) +void vloxseg3ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u64m2_m))) +void vloxseg3ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i64m1))) +void vloxseg4ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i64m1_m))) +void vloxseg4ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i64m2))) +void vloxseg4ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i64m2_m))) +void vloxseg4ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u64m1))) +void vloxseg4ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u64m1_m))) +void vloxseg4ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u64m2))) +void vloxseg4ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u64m2_m))) +void vloxseg4ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i64m1))) +void vloxseg5ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i64m1_m))) +void vloxseg5ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u64m1))) +void vloxseg5ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u64m1_m))) +void vloxseg5ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i64m1))) +void vloxseg6ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i64m1_m))) +void vloxseg6ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u64m1))) +void vloxseg6ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u64m1_m))) +void vloxseg6ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i64m1))) +void vloxseg7ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i64m1_m))) +void vloxseg7ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u64m1))) +void vloxseg7ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u64m1_m))) +void vloxseg7ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i64m1))) +void vloxseg8ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i64m1_m))) +void vloxseg8ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u64m1))) +void vloxseg8ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u64m1_m))) +void vloxseg8ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i64m1))) +void vloxseg2ei16(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i64m1_m))) +void vloxseg2ei16(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i64m2))) +void vloxseg2ei16(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i64m2_m))) +void vloxseg2ei16(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i64m4))) +void vloxseg2ei16(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i64m4_m))) +void vloxseg2ei16(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u64m1))) +void vloxseg2ei16(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u64m1_m))) +void vloxseg2ei16(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u64m2))) +void vloxseg2ei16(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u64m2_m))) +void vloxseg2ei16(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u64m4))) +void vloxseg2ei16(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u64m4_m))) +void vloxseg2ei16(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i64m1))) +void vloxseg3ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i64m1_m))) +void vloxseg3ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i64m2))) +void vloxseg3ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i64m2_m))) +void vloxseg3ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u64m1))) +void vloxseg3ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u64m1_m))) +void vloxseg3ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u64m2))) +void vloxseg3ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u64m2_m))) +void vloxseg3ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i64m1))) +void vloxseg4ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i64m1_m))) +void vloxseg4ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i64m2))) +void vloxseg4ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i64m2_m))) +void vloxseg4ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u64m1))) +void vloxseg4ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u64m1_m))) +void vloxseg4ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u64m2))) +void vloxseg4ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u64m2_m))) +void vloxseg4ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i64m1))) +void vloxseg5ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i64m1_m))) +void vloxseg5ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u64m1))) +void vloxseg5ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u64m1_m))) +void vloxseg5ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i64m1))) +void vloxseg6ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i64m1_m))) +void vloxseg6ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u64m1))) +void vloxseg6ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u64m1_m))) +void vloxseg6ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i64m1))) +void vloxseg7ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i64m1_m))) +void vloxseg7ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u64m1))) +void vloxseg7ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u64m1_m))) +void vloxseg7ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i64m1))) +void vloxseg8ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i64m1_m))) +void vloxseg8ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u64m1))) +void vloxseg8ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u64m1_m))) +void vloxseg8ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i64m1))) +void vloxseg2ei32(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i64m1_m))) +void vloxseg2ei32(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i64m2))) +void vloxseg2ei32(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i64m2_m))) +void vloxseg2ei32(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i64m4))) +void vloxseg2ei32(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i64m4_m))) +void vloxseg2ei32(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u64m1))) +void vloxseg2ei32(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u64m1_m))) +void vloxseg2ei32(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u64m2))) +void vloxseg2ei32(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u64m2_m))) +void vloxseg2ei32(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u64m4))) +void vloxseg2ei32(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u64m4_m))) +void vloxseg2ei32(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i64m1))) +void vloxseg3ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i64m1_m))) +void vloxseg3ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i64m2))) +void vloxseg3ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i64m2_m))) +void vloxseg3ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u64m1))) +void vloxseg3ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u64m1_m))) +void vloxseg3ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u64m2))) +void vloxseg3ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u64m2_m))) +void vloxseg3ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i64m1))) +void vloxseg4ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i64m1_m))) +void vloxseg4ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i64m2))) +void vloxseg4ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i64m2_m))) +void vloxseg4ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u64m1))) +void vloxseg4ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u64m1_m))) +void vloxseg4ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u64m2))) +void vloxseg4ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u64m2_m))) +void vloxseg4ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i64m1))) +void vloxseg5ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i64m1_m))) +void vloxseg5ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u64m1))) +void vloxseg5ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u64m1_m))) +void vloxseg5ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i64m1))) +void vloxseg6ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i64m1_m))) +void vloxseg6ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u64m1))) +void vloxseg6ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u64m1_m))) +void vloxseg6ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i64m1))) +void vloxseg7ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i64m1_m))) +void vloxseg7ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u64m1))) +void vloxseg7ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u64m1_m))) +void vloxseg7ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i64m1))) +void vloxseg8ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i64m1_m))) +void vloxseg8ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u64m1))) +void vloxseg8ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u64m1_m))) +void vloxseg8ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i64m1))) +void vloxseg2ei64(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i64m1_m))) +void vloxseg2ei64(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i64m2))) +void vloxseg2ei64(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i64m2_m))) +void vloxseg2ei64(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i64m4))) +void vloxseg2ei64(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i64m4_m))) +void vloxseg2ei64(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u64m1))) +void vloxseg2ei64(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u64m1_m))) +void vloxseg2ei64(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u64m2))) +void vloxseg2ei64(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u64m2_m))) +void vloxseg2ei64(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u64m4))) +void vloxseg2ei64(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u64m4_m))) +void vloxseg2ei64(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i64m1))) +void vloxseg3ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i64m1_m))) +void vloxseg3ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i64m2))) +void vloxseg3ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i64m2_m))) +void vloxseg3ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u64m1))) +void vloxseg3ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u64m1_m))) +void vloxseg3ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u64m2))) +void vloxseg3ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u64m2_m))) +void vloxseg3ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i64m1))) +void vloxseg4ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i64m1_m))) +void vloxseg4ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i64m2))) +void vloxseg4ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i64m2_m))) +void vloxseg4ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u64m1))) +void vloxseg4ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u64m1_m))) +void vloxseg4ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u64m2))) +void vloxseg4ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u64m2_m))) +void vloxseg4ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i64m1))) +void vloxseg5ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i64m1_m))) +void vloxseg5ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u64m1))) +void vloxseg5ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u64m1_m))) +void vloxseg5ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i64m1))) +void vloxseg6ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i64m1_m))) +void vloxseg6ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u64m1))) +void vloxseg6ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u64m1_m))) +void vloxseg6ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i64m1))) +void vloxseg7ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i64m1_m))) +void vloxseg7ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u64m1))) +void vloxseg7ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u64m1_m))) +void vloxseg7ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i64m1))) +void vloxseg8ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i64m1_m))) +void vloxseg8ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u64m1))) +void vloxseg8ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u64m1_m))) +void vloxseg8ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8m1))) +void vsseg3e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8m1_m))) +void vsseg3e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8m2))) +void vsseg3e8(int8_t * op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8m2_m))) +void vsseg3e8(vbool4_t op0, int8_t * op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8mf2))) +void vsseg3e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8mf2_m))) +void vsseg3e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8mf4))) +void vsseg3e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8mf4_m))) +void vsseg3e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8mf8))) +void vsseg3e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8mf8_m))) +void vsseg3e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8m1))) +void vsseg3e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8m1_m))) +void vsseg3e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8m2))) +void vsseg3e8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8m2_m))) +void vsseg3e8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8mf2))) +void vsseg3e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8mf2_m))) +void vsseg3e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8mf4))) +void vsseg3e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8mf4_m))) +void vsseg3e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8mf8))) +void vsseg3e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8mf8_m))) +void vsseg3e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8m1))) +void vsseg4e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8m1_m))) +void vsseg4e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8m2))) +void vsseg4e8(int8_t * op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8m2_m))) +void vsseg4e8(vbool4_t op0, int8_t * op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8mf2))) +void vsseg4e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8mf2_m))) +void vsseg4e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8mf4))) +void vsseg4e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8mf4_m))) +void vsseg4e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8mf8))) +void vsseg4e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8mf8_m))) +void vsseg4e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8m1))) +void vsseg4e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8m1_m))) +void vsseg4e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8m2))) +void vsseg4e8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8m2_m))) +void vsseg4e8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8mf2))) +void vsseg4e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8mf2_m))) +void vsseg4e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8mf4))) +void vsseg4e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8mf4_m))) +void vsseg4e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8mf8))) +void vsseg4e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8mf8_m))) +void vsseg4e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8m1))) +void vsseg5e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8m1_m))) +void vsseg5e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8mf2))) +void vsseg5e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8mf2_m))) +void vsseg5e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8mf4))) +void vsseg5e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8mf4_m))) +void vsseg5e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8mf8))) +void vsseg5e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8mf8_m))) +void vsseg5e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8m1))) +void vsseg5e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8m1_m))) +void vsseg5e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8mf2))) +void vsseg5e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8mf2_m))) +void vsseg5e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8mf4))) +void vsseg5e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8mf4_m))) +void vsseg5e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8mf8))) +void vsseg5e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8mf8_m))) +void vsseg5e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8m1))) +void vsseg6e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8m1_m))) +void vsseg6e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8mf2))) +void vsseg6e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8mf2_m))) +void vsseg6e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8mf4))) +void vsseg6e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8mf4_m))) +void vsseg6e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8mf8))) +void vsseg6e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8mf8_m))) +void vsseg6e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8m1))) +void vsseg6e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8m1_m))) +void vsseg6e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8mf2))) +void vsseg6e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8mf2_m))) +void vsseg6e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8mf4))) +void vsseg6e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8mf4_m))) +void vsseg6e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8mf8))) +void vsseg6e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8mf8_m))) +void vsseg6e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8m1))) +void vsseg7e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8m1_m))) +void vsseg7e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8mf2))) +void vsseg7e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8mf2_m))) +void vsseg7e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8mf4))) +void vsseg7e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8mf4_m))) +void vsseg7e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8mf8))) +void vsseg7e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8mf8_m))) +void vsseg7e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8m1))) +void vsseg7e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8m1_m))) +void vsseg7e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8mf2))) +void vsseg7e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8mf2_m))) +void vsseg7e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8mf4))) +void vsseg7e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8mf4_m))) +void vsseg7e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8mf8))) +void vsseg7e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8mf8_m))) +void vsseg7e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8m1))) +void vsseg8e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8m1_m))) +void vsseg8e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8mf2))) +void vsseg8e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8mf2_m))) +void vsseg8e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8mf4))) +void vsseg8e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8mf4_m))) +void vsseg8e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8mf8))) +void vsseg8e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8mf8_m))) +void vsseg8e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8m1))) +void vsseg8e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8m1_m))) +void vsseg8e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8mf2))) +void vsseg8e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8mf2_m))) +void vsseg8e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8mf4))) +void vsseg8e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8mf4_m))) +void vsseg8e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8mf8))) +void vsseg8e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8mf8_m))) +void vsseg8e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16m1))) +void vsseg2e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16m1_m))) +void vsseg2e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16m2))) +void vsseg2e16(int16_t * op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16m2_m))) +void vsseg2e16(vbool8_t op0, int16_t * op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16m4))) +void vsseg2e16(int16_t * op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16m4_m))) +void vsseg2e16(vbool4_t op0, int16_t * op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16mf2))) +void vsseg2e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16mf2_m))) +void vsseg2e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16mf4))) +void vsseg2e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16mf4_m))) +void vsseg2e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16m1))) +void vsseg2e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16m1_m))) +void vsseg2e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16m2))) +void vsseg2e16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16m2_m))) +void vsseg2e16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16m4))) +void vsseg2e16(uint16_t * op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16m4_m))) +void vsseg2e16(vbool4_t op0, uint16_t * op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16mf2))) +void vsseg2e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16mf2_m))) +void vsseg2e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16mf4))) +void vsseg2e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16mf4_m))) +void vsseg2e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16m1))) +void vsseg3e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16m1_m))) +void vsseg3e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16m2))) +void vsseg3e16(int16_t * op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16m2_m))) +void vsseg3e16(vbool8_t op0, int16_t * op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16mf2))) +void vsseg3e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16mf2_m))) +void vsseg3e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16mf4))) +void vsseg3e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16mf4_m))) +void vsseg3e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16m1))) +void vsseg3e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16m1_m))) +void vsseg3e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16m2))) +void vsseg3e16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16m2_m))) +void vsseg3e16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16mf2))) +void vsseg3e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16mf2_m))) +void vsseg3e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16mf4))) +void vsseg3e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16mf4_m))) +void vsseg3e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16m1))) +void vsseg4e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16m1_m))) +void vsseg4e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16m2))) +void vsseg4e16(int16_t * op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16m2_m))) +void vsseg4e16(vbool8_t op0, int16_t * op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16mf2))) +void vsseg4e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16mf2_m))) +void vsseg4e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16mf4))) +void vsseg4e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16mf4_m))) +void vsseg4e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16m1))) +void vsseg4e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16m1_m))) +void vsseg4e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16m2))) +void vsseg4e16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16m2_m))) +void vsseg4e16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16mf2))) +void vsseg4e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16mf2_m))) +void vsseg4e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16mf4))) +void vsseg4e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16mf4_m))) +void vsseg4e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_i16m1))) +void vsseg5e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_i16m1_m))) +void vsseg5e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_i16mf2))) +void vsseg5e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_i16mf2_m))) +void vsseg5e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_i16mf4))) +void vsseg5e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_i16mf4_m))) +void vsseg5e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_u16m1))) +void vsseg5e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_u16m1_m))) +void vsseg5e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_u16mf2))) +void vsseg5e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_u16mf2_m))) +void vsseg5e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_u16mf4))) +void vsseg5e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_u16mf4_m))) +void vsseg5e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_i16m1))) +void vsseg6e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_i16m1_m))) +void vsseg6e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_i16mf2))) +void vsseg6e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_i16mf2_m))) +void vsseg6e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_i16mf4))) +void vsseg6e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_i16mf4_m))) +void vsseg6e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_u16m1))) +void vsseg6e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_u16m1_m))) +void vsseg6e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_u16mf2))) +void vsseg6e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_u16mf2_m))) +void vsseg6e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_u16mf4))) +void vsseg6e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_u16mf4_m))) +void vsseg6e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_i16m1))) +void vsseg7e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_i16m1_m))) +void vsseg7e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_i16mf2))) +void vsseg7e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_i16mf2_m))) +void vsseg7e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_i16mf4))) +void vsseg7e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_i16mf4_m))) +void vsseg7e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_u16m1))) +void vsseg7e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_u16m1_m))) +void vsseg7e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_u16mf2))) +void vsseg7e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_u16mf2_m))) +void vsseg7e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_u16mf4))) +void vsseg7e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_u16mf4_m))) +void vsseg7e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_i16m1))) +void vsseg8e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_i16m1_m))) +void vsseg8e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_i16mf2))) +void vsseg8e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_i16mf2_m))) +void vsseg8e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_i16mf4))) +void vsseg8e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_i16mf4_m))) +void vsseg8e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_u16m1))) +void vsseg8e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_u16m1_m))) +void vsseg8e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_u16mf2))) +void vsseg8e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_u16mf2_m))) +void vsseg8e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_u16mf4))) +void vsseg8e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_u16mf4_m))) +void vsseg8e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32m1))) +void vsseg2e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32m1_m))) +void vsseg2e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32m2))) +void vsseg2e32(int32_t * op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32m2_m))) +void vsseg2e32(vbool16_t op0, int32_t * op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32m4))) +void vsseg2e32(int32_t * op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32m4_m))) +void vsseg2e32(vbool8_t op0, int32_t * op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32mf2))) +void vsseg2e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32mf2_m))) +void vsseg2e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32m1))) +void vsseg2e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32m1_m))) +void vsseg2e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32m2))) +void vsseg2e32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32m2_m))) +void vsseg2e32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32m4))) +void vsseg2e32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32m4_m))) +void vsseg2e32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32mf2))) +void vsseg2e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32mf2_m))) +void vsseg2e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_i32m1))) +void vsseg3e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_i32m1_m))) +void vsseg3e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_i32m2))) +void vsseg3e32(int32_t * op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_i32m2_m))) +void vsseg3e32(vbool16_t op0, int32_t * op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_i32mf2))) +void vsseg3e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_i32mf2_m))) +void vsseg3e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_u32m1))) +void vsseg3e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_u32m1_m))) +void vsseg3e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_u32m2))) +void vsseg3e32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_u32m2_m))) +void vsseg3e32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_u32mf2))) +void vsseg3e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_u32mf2_m))) +void vsseg3e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_i32m1))) +void vsseg4e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_i32m1_m))) +void vsseg4e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_i32m2))) +void vsseg4e32(int32_t * op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_i32m2_m))) +void vsseg4e32(vbool16_t op0, int32_t * op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_i32mf2))) +void vsseg4e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_i32mf2_m))) +void vsseg4e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_u32m1))) +void vsseg4e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_u32m1_m))) +void vsseg4e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_u32m2))) +void vsseg4e32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_u32m2_m))) +void vsseg4e32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_u32mf2))) +void vsseg4e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_u32mf2_m))) +void vsseg4e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_i32m1))) +void vsseg5e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_i32m1_m))) +void vsseg5e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_i32mf2))) +void vsseg5e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_i32mf2_m))) +void vsseg5e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_u32m1))) +void vsseg5e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_u32m1_m))) +void vsseg5e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_u32mf2))) +void vsseg5e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_u32mf2_m))) +void vsseg5e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_i32m1))) +void vsseg6e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_i32m1_m))) +void vsseg6e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_i32mf2))) +void vsseg6e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_i32mf2_m))) +void vsseg6e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_u32m1))) +void vsseg6e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_u32m1_m))) +void vsseg6e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_u32mf2))) +void vsseg6e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_u32mf2_m))) +void vsseg6e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_i32m1))) +void vsseg7e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_i32m1_m))) +void vsseg7e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_i32mf2))) +void vsseg7e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_i32mf2_m))) +void vsseg7e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_u32m1))) +void vsseg7e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_u32m1_m))) +void vsseg7e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_u32mf2))) +void vsseg7e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_u32mf2_m))) +void vsseg7e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_i32m1))) +void vsseg8e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_i32m1_m))) +void vsseg8e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_i32mf2))) +void vsseg8e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_i32mf2_m))) +void vsseg8e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_u32m1))) +void vsseg8e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_u32m1_m))) +void vsseg8e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_u32mf2))) +void vsseg8e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_u32mf2_m))) +void vsseg8e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_i64m1))) +void vsseg2e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_i64m1_m))) +void vsseg2e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_i64m2))) +void vsseg2e64(int64_t * op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_i64m2_m))) +void vsseg2e64(vbool32_t op0, int64_t * op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_i64m4))) +void vsseg2e64(int64_t * op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_i64m4_m))) +void vsseg2e64(vbool16_t op0, int64_t * op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_u64m1))) +void vsseg2e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_u64m1_m))) +void vsseg2e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_u64m2))) +void vsseg2e64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_u64m2_m))) +void vsseg2e64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_u64m4))) +void vsseg2e64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_u64m4_m))) +void vsseg2e64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_i64m1))) +void vsseg3e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_i64m1_m))) +void vsseg3e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_i64m2))) +void vsseg3e64(int64_t * op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_i64m2_m))) +void vsseg3e64(vbool32_t op0, int64_t * op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_u64m1))) +void vsseg3e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_u64m1_m))) +void vsseg3e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_u64m2))) +void vsseg3e64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_u64m2_m))) +void vsseg3e64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_i64m1))) +void vsseg4e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_i64m1_m))) +void vsseg4e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_i64m2))) +void vsseg4e64(int64_t * op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_i64m2_m))) +void vsseg4e64(vbool32_t op0, int64_t * op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_u64m1))) +void vsseg4e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_u64m1_m))) +void vsseg4e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_u64m2))) +void vsseg4e64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_u64m2_m))) +void vsseg4e64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e64_v_i64m1))) +void vsseg5e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e64_v_i64m1_m))) +void vsseg5e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e64_v_u64m1))) +void vsseg5e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e64_v_u64m1_m))) +void vsseg5e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e64_v_i64m1))) +void vsseg6e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e64_v_i64m1_m))) +void vsseg6e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e64_v_u64m1))) +void vsseg6e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e64_v_u64m1_m))) +void vsseg6e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e64_v_i64m1))) +void vsseg7e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e64_v_i64m1_m))) +void vsseg7e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e64_v_u64m1))) +void vsseg7e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e64_v_u64m1_m))) +void vsseg7e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e64_v_i64m1))) +void vsseg8e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e64_v_i64m1_m))) +void vsseg8e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e64_v_u64m1))) +void vsseg8e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e64_v_u64m1_m))) +void vsseg8e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8m1))) +void vssseg3e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8m1_m))) +void vssseg3e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8m2))) +void vssseg3e8(int8_t * op0, ptrdiff_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8m2_m))) +void vssseg3e8(vbool4_t op0, int8_t * op1, ptrdiff_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8mf2))) +void vssseg3e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8mf2_m))) +void vssseg3e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8mf4))) +void vssseg3e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8mf4_m))) +void vssseg3e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8mf8))) +void vssseg3e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8mf8_m))) +void vssseg3e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8m1))) +void vssseg3e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8m1_m))) +void vssseg3e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8m2))) +void vssseg3e8(uint8_t * op0, ptrdiff_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8m2_m))) +void vssseg3e8(vbool4_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8mf2))) +void vssseg3e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8mf2_m))) +void vssseg3e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8mf4))) +void vssseg3e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8mf4_m))) +void vssseg3e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8mf8))) +void vssseg3e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8mf8_m))) +void vssseg3e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8m1))) +void vssseg4e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8m1_m))) +void vssseg4e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8m2))) +void vssseg4e8(int8_t * op0, ptrdiff_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8m2_m))) +void vssseg4e8(vbool4_t op0, int8_t * op1, ptrdiff_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8mf2))) +void vssseg4e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8mf2_m))) +void vssseg4e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8mf4))) +void vssseg4e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8mf4_m))) +void vssseg4e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8mf8))) +void vssseg4e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8mf8_m))) +void vssseg4e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8m1))) +void vssseg4e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8m1_m))) +void vssseg4e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8m2))) +void vssseg4e8(uint8_t * op0, ptrdiff_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8m2_m))) +void vssseg4e8(vbool4_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8mf2))) +void vssseg4e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8mf2_m))) +void vssseg4e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8mf4))) +void vssseg4e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8mf4_m))) +void vssseg4e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8mf8))) +void vssseg4e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8mf8_m))) +void vssseg4e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8m1))) +void vssseg5e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8m1_m))) +void vssseg5e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8mf2))) +void vssseg5e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8mf2_m))) +void vssseg5e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8mf4))) +void vssseg5e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8mf4_m))) +void vssseg5e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8mf8))) +void vssseg5e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8mf8_m))) +void vssseg5e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8m1))) +void vssseg5e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8m1_m))) +void vssseg5e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8mf2))) +void vssseg5e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8mf2_m))) +void vssseg5e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8mf4))) +void vssseg5e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8mf4_m))) +void vssseg5e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8mf8))) +void vssseg5e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8mf8_m))) +void vssseg5e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8m1))) +void vssseg6e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8m1_m))) +void vssseg6e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8mf2))) +void vssseg6e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8mf2_m))) +void vssseg6e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8mf4))) +void vssseg6e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8mf4_m))) +void vssseg6e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8mf8))) +void vssseg6e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8mf8_m))) +void vssseg6e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8m1))) +void vssseg6e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8m1_m))) +void vssseg6e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8mf2))) +void vssseg6e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8mf2_m))) +void vssseg6e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8mf4))) +void vssseg6e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8mf4_m))) +void vssseg6e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8mf8))) +void vssseg6e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8mf8_m))) +void vssseg6e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8m1))) +void vssseg7e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8m1_m))) +void vssseg7e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8mf2))) +void vssseg7e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8mf2_m))) +void vssseg7e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8mf4))) +void vssseg7e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8mf4_m))) +void vssseg7e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8mf8))) +void vssseg7e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8mf8_m))) +void vssseg7e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8m1))) +void vssseg7e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8m1_m))) +void vssseg7e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8mf2))) +void vssseg7e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8mf2_m))) +void vssseg7e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8mf4))) +void vssseg7e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8mf4_m))) +void vssseg7e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8mf8))) +void vssseg7e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8mf8_m))) +void vssseg7e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8m1))) +void vssseg8e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8m1_m))) +void vssseg8e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8mf2))) +void vssseg8e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8mf2_m))) +void vssseg8e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8mf4))) +void vssseg8e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8mf4_m))) +void vssseg8e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8mf8))) +void vssseg8e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8mf8_m))) +void vssseg8e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8m1))) +void vssseg8e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8m1_m))) +void vssseg8e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8mf2))) +void vssseg8e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8mf2_m))) +void vssseg8e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8mf4))) +void vssseg8e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8mf4_m))) +void vssseg8e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8mf8))) +void vssseg8e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8mf8_m))) +void vssseg8e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16m1))) +void vssseg2e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16m1_m))) +void vssseg2e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16m2))) +void vssseg2e16(int16_t * op0, ptrdiff_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16m2_m))) +void vssseg2e16(vbool8_t op0, int16_t * op1, ptrdiff_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16m4))) +void vssseg2e16(int16_t * op0, ptrdiff_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16m4_m))) +void vssseg2e16(vbool4_t op0, int16_t * op1, ptrdiff_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16mf2))) +void vssseg2e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16mf2_m))) +void vssseg2e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16mf4))) +void vssseg2e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16mf4_m))) +void vssseg2e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16m1))) +void vssseg2e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16m1_m))) +void vssseg2e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16m2))) +void vssseg2e16(uint16_t * op0, ptrdiff_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16m2_m))) +void vssseg2e16(vbool8_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16m4))) +void vssseg2e16(uint16_t * op0, ptrdiff_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16m4_m))) +void vssseg2e16(vbool4_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16mf2))) +void vssseg2e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16mf2_m))) +void vssseg2e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16mf4))) +void vssseg2e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16mf4_m))) +void vssseg2e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16m1))) +void vssseg3e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16m1_m))) +void vssseg3e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16m2))) +void vssseg3e16(int16_t * op0, ptrdiff_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16m2_m))) +void vssseg3e16(vbool8_t op0, int16_t * op1, ptrdiff_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16mf2))) +void vssseg3e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16mf2_m))) +void vssseg3e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16mf4))) +void vssseg3e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16mf4_m))) +void vssseg3e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16m1))) +void vssseg3e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16m1_m))) +void vssseg3e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16m2))) +void vssseg3e16(uint16_t * op0, ptrdiff_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16m2_m))) +void vssseg3e16(vbool8_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16mf2))) +void vssseg3e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16mf2_m))) +void vssseg3e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16mf4))) +void vssseg3e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16mf4_m))) +void vssseg3e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16m1))) +void vssseg4e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16m1_m))) +void vssseg4e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16m2))) +void vssseg4e16(int16_t * op0, ptrdiff_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16m2_m))) +void vssseg4e16(vbool8_t op0, int16_t * op1, ptrdiff_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16mf2))) +void vssseg4e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16mf2_m))) +void vssseg4e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16mf4))) +void vssseg4e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16mf4_m))) +void vssseg4e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16m1))) +void vssseg4e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16m1_m))) +void vssseg4e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16m2))) +void vssseg4e16(uint16_t * op0, ptrdiff_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16m2_m))) +void vssseg4e16(vbool8_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16mf2))) +void vssseg4e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16mf2_m))) +void vssseg4e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16mf4))) +void vssseg4e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16mf4_m))) +void vssseg4e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_i16m1))) +void vssseg5e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_i16m1_m))) +void vssseg5e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_i16mf2))) +void vssseg5e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_i16mf2_m))) +void vssseg5e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_i16mf4))) +void vssseg5e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_i16mf4_m))) +void vssseg5e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_u16m1))) +void vssseg5e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_u16m1_m))) +void vssseg5e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_u16mf2))) +void vssseg5e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_u16mf2_m))) +void vssseg5e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_u16mf4))) +void vssseg5e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_u16mf4_m))) +void vssseg5e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_i16m1))) +void vssseg6e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_i16m1_m))) +void vssseg6e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_i16mf2))) +void vssseg6e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_i16mf2_m))) +void vssseg6e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_i16mf4))) +void vssseg6e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_i16mf4_m))) +void vssseg6e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_u16m1))) +void vssseg6e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_u16m1_m))) +void vssseg6e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_u16mf2))) +void vssseg6e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_u16mf2_m))) +void vssseg6e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_u16mf4))) +void vssseg6e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_u16mf4_m))) +void vssseg6e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_i16m1))) +void vssseg7e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_i16m1_m))) +void vssseg7e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_i16mf2))) +void vssseg7e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_i16mf2_m))) +void vssseg7e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_i16mf4))) +void vssseg7e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_i16mf4_m))) +void vssseg7e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_u16m1))) +void vssseg7e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_u16m1_m))) +void vssseg7e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_u16mf2))) +void vssseg7e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_u16mf2_m))) +void vssseg7e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_u16mf4))) +void vssseg7e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_u16mf4_m))) +void vssseg7e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_i16m1))) +void vssseg8e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_i16m1_m))) +void vssseg8e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_i16mf2))) +void vssseg8e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_i16mf2_m))) +void vssseg8e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_i16mf4))) +void vssseg8e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_i16mf4_m))) +void vssseg8e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_u16m1))) +void vssseg8e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_u16m1_m))) +void vssseg8e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_u16mf2))) +void vssseg8e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_u16mf2_m))) +void vssseg8e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_u16mf4))) +void vssseg8e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_u16mf4_m))) +void vssseg8e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32m1))) +void vssseg2e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32m1_m))) +void vssseg2e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32m2))) +void vssseg2e32(int32_t * op0, ptrdiff_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32m2_m))) +void vssseg2e32(vbool16_t op0, int32_t * op1, ptrdiff_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32m4))) +void vssseg2e32(int32_t * op0, ptrdiff_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32m4_m))) +void vssseg2e32(vbool8_t op0, int32_t * op1, ptrdiff_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32mf2))) +void vssseg2e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32mf2_m))) +void vssseg2e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32m1))) +void vssseg2e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32m1_m))) +void vssseg2e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32m2))) +void vssseg2e32(uint32_t * op0, ptrdiff_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32m2_m))) +void vssseg2e32(vbool16_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32m4))) +void vssseg2e32(uint32_t * op0, ptrdiff_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32m4_m))) +void vssseg2e32(vbool8_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32mf2))) +void vssseg2e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32mf2_m))) +void vssseg2e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_i32m1))) +void vssseg3e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_i32m1_m))) +void vssseg3e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_i32m2))) +void vssseg3e32(int32_t * op0, ptrdiff_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_i32m2_m))) +void vssseg3e32(vbool16_t op0, int32_t * op1, ptrdiff_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_i32mf2))) +void vssseg3e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_i32mf2_m))) +void vssseg3e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_u32m1))) +void vssseg3e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_u32m1_m))) +void vssseg3e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_u32m2))) +void vssseg3e32(uint32_t * op0, ptrdiff_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_u32m2_m))) +void vssseg3e32(vbool16_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_u32mf2))) +void vssseg3e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_u32mf2_m))) +void vssseg3e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_i32m1))) +void vssseg4e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_i32m1_m))) +void vssseg4e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_i32m2))) +void vssseg4e32(int32_t * op0, ptrdiff_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_i32m2_m))) +void vssseg4e32(vbool16_t op0, int32_t * op1, ptrdiff_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_i32mf2))) +void vssseg4e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_i32mf2_m))) +void vssseg4e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_u32m1))) +void vssseg4e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_u32m1_m))) +void vssseg4e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_u32m2))) +void vssseg4e32(uint32_t * op0, ptrdiff_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_u32m2_m))) +void vssseg4e32(vbool16_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_u32mf2))) +void vssseg4e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_u32mf2_m))) +void vssseg4e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_i32m1))) +void vssseg5e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_i32m1_m))) +void vssseg5e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_i32mf2))) +void vssseg5e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_i32mf2_m))) +void vssseg5e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_u32m1))) +void vssseg5e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_u32m1_m))) +void vssseg5e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_u32mf2))) +void vssseg5e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_u32mf2_m))) +void vssseg5e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_i32m1))) +void vssseg6e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_i32m1_m))) +void vssseg6e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_i32mf2))) +void vssseg6e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_i32mf2_m))) +void vssseg6e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_u32m1))) +void vssseg6e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_u32m1_m))) +void vssseg6e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_u32mf2))) +void vssseg6e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_u32mf2_m))) +void vssseg6e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_i32m1))) +void vssseg7e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_i32m1_m))) +void vssseg7e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_i32mf2))) +void vssseg7e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_i32mf2_m))) +void vssseg7e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_u32m1))) +void vssseg7e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_u32m1_m))) +void vssseg7e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_u32mf2))) +void vssseg7e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_u32mf2_m))) +void vssseg7e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_i32m1))) +void vssseg8e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_i32m1_m))) +void vssseg8e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_i32mf2))) +void vssseg8e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_i32mf2_m))) +void vssseg8e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_u32m1))) +void vssseg8e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_u32m1_m))) +void vssseg8e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_u32mf2))) +void vssseg8e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_u32mf2_m))) +void vssseg8e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_i64m1))) +void vssseg2e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_i64m1_m))) +void vssseg2e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_i64m2))) +void vssseg2e64(int64_t * op0, ptrdiff_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_i64m2_m))) +void vssseg2e64(vbool32_t op0, int64_t * op1, ptrdiff_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_i64m4))) +void vssseg2e64(int64_t * op0, ptrdiff_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_i64m4_m))) +void vssseg2e64(vbool16_t op0, int64_t * op1, ptrdiff_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_u64m1))) +void vssseg2e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_u64m1_m))) +void vssseg2e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_u64m2))) +void vssseg2e64(uint64_t * op0, ptrdiff_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_u64m2_m))) +void vssseg2e64(vbool32_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_u64m4))) +void vssseg2e64(uint64_t * op0, ptrdiff_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_u64m4_m))) +void vssseg2e64(vbool16_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_i64m1))) +void vssseg3e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_i64m1_m))) +void vssseg3e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_i64m2))) +void vssseg3e64(int64_t * op0, ptrdiff_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_i64m2_m))) +void vssseg3e64(vbool32_t op0, int64_t * op1, ptrdiff_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_u64m1))) +void vssseg3e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_u64m1_m))) +void vssseg3e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_u64m2))) +void vssseg3e64(uint64_t * op0, ptrdiff_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_u64m2_m))) +void vssseg3e64(vbool32_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_i64m1))) +void vssseg4e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_i64m1_m))) +void vssseg4e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_i64m2))) +void vssseg4e64(int64_t * op0, ptrdiff_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_i64m2_m))) +void vssseg4e64(vbool32_t op0, int64_t * op1, ptrdiff_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_u64m1))) +void vssseg4e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_u64m1_m))) +void vssseg4e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_u64m2))) +void vssseg4e64(uint64_t * op0, ptrdiff_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_u64m2_m))) +void vssseg4e64(vbool32_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e64_v_i64m1))) +void vssseg5e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e64_v_i64m1_m))) +void vssseg5e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e64_v_u64m1))) +void vssseg5e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e64_v_u64m1_m))) +void vssseg5e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e64_v_i64m1))) +void vssseg6e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e64_v_i64m1_m))) +void vssseg6e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e64_v_u64m1))) +void vssseg6e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e64_v_u64m1_m))) +void vssseg6e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e64_v_i64m1))) +void vssseg7e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e64_v_i64m1_m))) +void vssseg7e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e64_v_u64m1))) +void vssseg7e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e64_v_u64m1_m))) +void vssseg7e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e64_v_i64m1))) +void vssseg8e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e64_v_i64m1_m))) +void vssseg8e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e64_v_u64m1))) +void vssseg8e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e64_v_u64m1_m))) +void vssseg8e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8m1))) +void vsuxseg3ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8m1_m))) +void vsuxseg3ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8m2))) +void vsuxseg3ei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8m2_m))) +void vsuxseg3ei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8mf2))) +void vsuxseg3ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8mf2_m))) +void vsuxseg3ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8mf4))) +void vsuxseg3ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8mf4_m))) +void vsuxseg3ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8mf8))) +void vsuxseg3ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8mf8_m))) +void vsuxseg3ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8m1))) +void vsuxseg3ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8m1_m))) +void vsuxseg3ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8m2))) +void vsuxseg3ei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8m2_m))) +void vsuxseg3ei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8mf2))) +void vsuxseg3ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8mf2_m))) +void vsuxseg3ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8mf4))) +void vsuxseg3ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8mf4_m))) +void vsuxseg3ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8mf8))) +void vsuxseg3ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8mf8_m))) +void vsuxseg3ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8m1))) +void vsuxseg4ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8m1_m))) +void vsuxseg4ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8m2))) +void vsuxseg4ei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8m2_m))) +void vsuxseg4ei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8mf2))) +void vsuxseg4ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8mf2_m))) +void vsuxseg4ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8mf4))) +void vsuxseg4ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8mf4_m))) +void vsuxseg4ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8mf8))) +void vsuxseg4ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8mf8_m))) +void vsuxseg4ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8m1))) +void vsuxseg4ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8m1_m))) +void vsuxseg4ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8m2))) +void vsuxseg4ei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8m2_m))) +void vsuxseg4ei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8mf2))) +void vsuxseg4ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8mf2_m))) +void vsuxseg4ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8mf4))) +void vsuxseg4ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8mf4_m))) +void vsuxseg4ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8mf8))) +void vsuxseg4ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8mf8_m))) +void vsuxseg4ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8m1))) +void vsuxseg5ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8m1_m))) +void vsuxseg5ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8mf2))) +void vsuxseg5ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8mf2_m))) +void vsuxseg5ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8mf4))) +void vsuxseg5ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8mf4_m))) +void vsuxseg5ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8mf8))) +void vsuxseg5ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8mf8_m))) +void vsuxseg5ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8m1))) +void vsuxseg5ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8m1_m))) +void vsuxseg5ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8mf2))) +void vsuxseg5ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8mf2_m))) +void vsuxseg5ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8mf4))) +void vsuxseg5ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8mf4_m))) +void vsuxseg5ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8mf8))) +void vsuxseg5ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8mf8_m))) +void vsuxseg5ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8m1))) +void vsuxseg6ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8m1_m))) +void vsuxseg6ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8mf2))) +void vsuxseg6ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8mf2_m))) +void vsuxseg6ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8mf4))) +void vsuxseg6ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8mf4_m))) +void vsuxseg6ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8mf8))) +void vsuxseg6ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8mf8_m))) +void vsuxseg6ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8m1))) +void vsuxseg6ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8m1_m))) +void vsuxseg6ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8mf2))) +void vsuxseg6ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8mf2_m))) +void vsuxseg6ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8mf4))) +void vsuxseg6ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8mf4_m))) +void vsuxseg6ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8mf8))) +void vsuxseg6ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8mf8_m))) +void vsuxseg6ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8m1))) +void vsuxseg7ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8m1_m))) +void vsuxseg7ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8mf2))) +void vsuxseg7ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8mf2_m))) +void vsuxseg7ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8mf4))) +void vsuxseg7ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8mf4_m))) +void vsuxseg7ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8mf8))) +void vsuxseg7ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8mf8_m))) +void vsuxseg7ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8m1))) +void vsuxseg7ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8m1_m))) +void vsuxseg7ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8mf2))) +void vsuxseg7ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8mf2_m))) +void vsuxseg7ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8mf4))) +void vsuxseg7ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8mf4_m))) +void vsuxseg7ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8mf8))) +void vsuxseg7ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8mf8_m))) +void vsuxseg7ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8m1))) +void vsuxseg8ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8m1_m))) +void vsuxseg8ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8mf2))) +void vsuxseg8ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8mf2_m))) +void vsuxseg8ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8mf4))) +void vsuxseg8ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8mf4_m))) +void vsuxseg8ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8mf8))) +void vsuxseg8ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8mf8_m))) +void vsuxseg8ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8m1))) +void vsuxseg8ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8m1_m))) +void vsuxseg8ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8mf2))) +void vsuxseg8ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8mf2_m))) +void vsuxseg8ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8mf4))) +void vsuxseg8ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8mf4_m))) +void vsuxseg8ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8mf8))) +void vsuxseg8ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8mf8_m))) +void vsuxseg8ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8m1))) +void vsuxseg2ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8m1_m))) +void vsuxseg2ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8m2))) +void vsuxseg2ei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8m2_m))) +void vsuxseg2ei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8m4))) +void vsuxseg2ei16(int8_t * op0, vuint16m8_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8m4_m))) +void vsuxseg2ei16(vbool2_t op0, int8_t * op1, vuint16m8_t op2, vint8m4_t op3, vint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8mf2))) +void vsuxseg2ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8mf2_m))) +void vsuxseg2ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8mf4))) +void vsuxseg2ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8mf4_m))) +void vsuxseg2ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8mf8))) +void vsuxseg2ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8mf8_m))) +void vsuxseg2ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8m1))) +void vsuxseg2ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8m1_m))) +void vsuxseg2ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8m2))) +void vsuxseg2ei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8m2_m))) +void vsuxseg2ei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8m4))) +void vsuxseg2ei16(uint8_t * op0, vuint16m8_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8m4_m))) +void vsuxseg2ei16(vbool2_t op0, uint8_t * op1, vuint16m8_t op2, vuint8m4_t op3, vuint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8mf2))) +void vsuxseg2ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8mf2_m))) +void vsuxseg2ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8mf4))) +void vsuxseg2ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8mf4_m))) +void vsuxseg2ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8mf8))) +void vsuxseg2ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8mf8_m))) +void vsuxseg2ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8m1))) +void vsuxseg3ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8m1_m))) +void vsuxseg3ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8m2))) +void vsuxseg3ei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8m2_m))) +void vsuxseg3ei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8mf2))) +void vsuxseg3ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8mf2_m))) +void vsuxseg3ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8mf4))) +void vsuxseg3ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8mf4_m))) +void vsuxseg3ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8mf8))) +void vsuxseg3ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8mf8_m))) +void vsuxseg3ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8m1))) +void vsuxseg3ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8m1_m))) +void vsuxseg3ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8m2))) +void vsuxseg3ei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8m2_m))) +void vsuxseg3ei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8mf2))) +void vsuxseg3ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8mf2_m))) +void vsuxseg3ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8mf4))) +void vsuxseg3ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8mf4_m))) +void vsuxseg3ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8mf8))) +void vsuxseg3ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8mf8_m))) +void vsuxseg3ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8m1))) +void vsuxseg4ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8m1_m))) +void vsuxseg4ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8m2))) +void vsuxseg4ei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8m2_m))) +void vsuxseg4ei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8mf2))) +void vsuxseg4ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8mf2_m))) +void vsuxseg4ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8mf4))) +void vsuxseg4ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8mf4_m))) +void vsuxseg4ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8mf8))) +void vsuxseg4ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8mf8_m))) +void vsuxseg4ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8m1))) +void vsuxseg4ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8m1_m))) +void vsuxseg4ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8m2))) +void vsuxseg4ei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8m2_m))) +void vsuxseg4ei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8mf2))) +void vsuxseg4ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8mf2_m))) +void vsuxseg4ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8mf4))) +void vsuxseg4ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8mf4_m))) +void vsuxseg4ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8mf8))) +void vsuxseg4ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8mf8_m))) +void vsuxseg4ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8m1))) +void vsuxseg5ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8m1_m))) +void vsuxseg5ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8mf2))) +void vsuxseg5ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8mf2_m))) +void vsuxseg5ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8mf4))) +void vsuxseg5ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8mf4_m))) +void vsuxseg5ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8mf8))) +void vsuxseg5ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8mf8_m))) +void vsuxseg5ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8m1))) +void vsuxseg5ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8m1_m))) +void vsuxseg5ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8mf2))) +void vsuxseg5ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8mf2_m))) +void vsuxseg5ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8mf4))) +void vsuxseg5ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8mf4_m))) +void vsuxseg5ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8mf8))) +void vsuxseg5ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8mf8_m))) +void vsuxseg5ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8m1))) +void vsuxseg6ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8m1_m))) +void vsuxseg6ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8mf2))) +void vsuxseg6ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8mf2_m))) +void vsuxseg6ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8mf4))) +void vsuxseg6ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8mf4_m))) +void vsuxseg6ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8mf8))) +void vsuxseg6ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8mf8_m))) +void vsuxseg6ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8m1))) +void vsuxseg6ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8m1_m))) +void vsuxseg6ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8mf2))) +void vsuxseg6ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8mf2_m))) +void vsuxseg6ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8mf4))) +void vsuxseg6ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8mf4_m))) +void vsuxseg6ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8mf8))) +void vsuxseg6ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8mf8_m))) +void vsuxseg6ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8m1))) +void vsuxseg7ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8m1_m))) +void vsuxseg7ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8mf2))) +void vsuxseg7ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8mf2_m))) +void vsuxseg7ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8mf4))) +void vsuxseg7ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8mf4_m))) +void vsuxseg7ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8mf8))) +void vsuxseg7ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8mf8_m))) +void vsuxseg7ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8m1))) +void vsuxseg7ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8m1_m))) +void vsuxseg7ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8mf2))) +void vsuxseg7ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8mf2_m))) +void vsuxseg7ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8mf4))) +void vsuxseg7ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8mf4_m))) +void vsuxseg7ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8mf8))) +void vsuxseg7ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8mf8_m))) +void vsuxseg7ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8m1))) +void vsuxseg8ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8m1_m))) +void vsuxseg8ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8mf2))) +void vsuxseg8ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8mf2_m))) +void vsuxseg8ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8mf4))) +void vsuxseg8ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8mf4_m))) +void vsuxseg8ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8mf8))) +void vsuxseg8ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8mf8_m))) +void vsuxseg8ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8m1))) +void vsuxseg8ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8m1_m))) +void vsuxseg8ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8mf2))) +void vsuxseg8ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8mf2_m))) +void vsuxseg8ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8mf4))) +void vsuxseg8ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8mf4_m))) +void vsuxseg8ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8mf8))) +void vsuxseg8ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8mf8_m))) +void vsuxseg8ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8m1))) +void vsuxseg2ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8m1_m))) +void vsuxseg2ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8m2))) +void vsuxseg2ei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8m2_m))) +void vsuxseg2ei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8mf2))) +void vsuxseg2ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8mf2_m))) +void vsuxseg2ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8mf4))) +void vsuxseg2ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8mf4_m))) +void vsuxseg2ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8mf8))) +void vsuxseg2ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8mf8_m))) +void vsuxseg2ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8m1))) +void vsuxseg2ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8m1_m))) +void vsuxseg2ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8m2))) +void vsuxseg2ei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8m2_m))) +void vsuxseg2ei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8mf2))) +void vsuxseg2ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8mf2_m))) +void vsuxseg2ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8mf4))) +void vsuxseg2ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8mf4_m))) +void vsuxseg2ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8mf8))) +void vsuxseg2ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8mf8_m))) +void vsuxseg2ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8m1))) +void vsuxseg3ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8m1_m))) +void vsuxseg3ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8m2))) +void vsuxseg3ei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8m2_m))) +void vsuxseg3ei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8mf2))) +void vsuxseg3ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8mf2_m))) +void vsuxseg3ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8mf4))) +void vsuxseg3ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8mf4_m))) +void vsuxseg3ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8mf8))) +void vsuxseg3ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8mf8_m))) +void vsuxseg3ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8m1))) +void vsuxseg3ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8m1_m))) +void vsuxseg3ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8m2))) +void vsuxseg3ei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8m2_m))) +void vsuxseg3ei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8mf2))) +void vsuxseg3ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8mf2_m))) +void vsuxseg3ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8mf4))) +void vsuxseg3ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8mf4_m))) +void vsuxseg3ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8mf8))) +void vsuxseg3ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8mf8_m))) +void vsuxseg3ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8m1))) +void vsuxseg4ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8m1_m))) +void vsuxseg4ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8m2))) +void vsuxseg4ei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8m2_m))) +void vsuxseg4ei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8mf2))) +void vsuxseg4ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8mf2_m))) +void vsuxseg4ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8mf4))) +void vsuxseg4ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8mf4_m))) +void vsuxseg4ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8mf8))) +void vsuxseg4ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8mf8_m))) +void vsuxseg4ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8m1))) +void vsuxseg4ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8m1_m))) +void vsuxseg4ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8m2))) +void vsuxseg4ei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8m2_m))) +void vsuxseg4ei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8mf2))) +void vsuxseg4ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8mf2_m))) +void vsuxseg4ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8mf4))) +void vsuxseg4ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8mf4_m))) +void vsuxseg4ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8mf8))) +void vsuxseg4ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8mf8_m))) +void vsuxseg4ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8m1))) +void vsuxseg5ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8m1_m))) +void vsuxseg5ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8mf2))) +void vsuxseg5ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8mf2_m))) +void vsuxseg5ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8mf4))) +void vsuxseg5ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8mf4_m))) +void vsuxseg5ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8mf8))) +void vsuxseg5ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8mf8_m))) +void vsuxseg5ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8m1))) +void vsuxseg5ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8m1_m))) +void vsuxseg5ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8mf2))) +void vsuxseg5ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8mf2_m))) +void vsuxseg5ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8mf4))) +void vsuxseg5ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8mf4_m))) +void vsuxseg5ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8mf8))) +void vsuxseg5ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8mf8_m))) +void vsuxseg5ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8m1))) +void vsuxseg6ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8m1_m))) +void vsuxseg6ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8mf2))) +void vsuxseg6ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8mf2_m))) +void vsuxseg6ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8mf4))) +void vsuxseg6ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8mf4_m))) +void vsuxseg6ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8mf8))) +void vsuxseg6ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8mf8_m))) +void vsuxseg6ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8m1))) +void vsuxseg6ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8m1_m))) +void vsuxseg6ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8mf2))) +void vsuxseg6ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8mf2_m))) +void vsuxseg6ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8mf4))) +void vsuxseg6ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8mf4_m))) +void vsuxseg6ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8mf8))) +void vsuxseg6ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8mf8_m))) +void vsuxseg6ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8m1))) +void vsuxseg7ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8m1_m))) +void vsuxseg7ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8mf2))) +void vsuxseg7ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8mf2_m))) +void vsuxseg7ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8mf4))) +void vsuxseg7ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8mf4_m))) +void vsuxseg7ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8mf8))) +void vsuxseg7ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8mf8_m))) +void vsuxseg7ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8m1))) +void vsuxseg7ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8m1_m))) +void vsuxseg7ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8mf2))) +void vsuxseg7ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8mf2_m))) +void vsuxseg7ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8mf4))) +void vsuxseg7ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8mf4_m))) +void vsuxseg7ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8mf8))) +void vsuxseg7ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8mf8_m))) +void vsuxseg7ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8m1))) +void vsuxseg8ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8m1_m))) +void vsuxseg8ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8mf2))) +void vsuxseg8ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8mf2_m))) +void vsuxseg8ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8mf4))) +void vsuxseg8ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8mf4_m))) +void vsuxseg8ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8mf8))) +void vsuxseg8ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8mf8_m))) +void vsuxseg8ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8m1))) +void vsuxseg8ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8m1_m))) +void vsuxseg8ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8mf2))) +void vsuxseg8ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8mf2_m))) +void vsuxseg8ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8mf4))) +void vsuxseg8ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8mf4_m))) +void vsuxseg8ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8mf8))) +void vsuxseg8ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8mf8_m))) +void vsuxseg8ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8m1))) +void vsuxseg2ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8m1_m))) +void vsuxseg2ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8mf2))) +void vsuxseg2ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8mf2_m))) +void vsuxseg2ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8mf4))) +void vsuxseg2ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8mf4_m))) +void vsuxseg2ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8mf8))) +void vsuxseg2ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8mf8_m))) +void vsuxseg2ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8m1))) +void vsuxseg2ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8m1_m))) +void vsuxseg2ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8mf2))) +void vsuxseg2ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8mf2_m))) +void vsuxseg2ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8mf4))) +void vsuxseg2ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8mf4_m))) +void vsuxseg2ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8mf8))) +void vsuxseg2ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8mf8_m))) +void vsuxseg2ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8m1))) +void vsuxseg3ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8m1_m))) +void vsuxseg3ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8mf2))) +void vsuxseg3ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8mf2_m))) +void vsuxseg3ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8mf4))) +void vsuxseg3ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8mf4_m))) +void vsuxseg3ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8mf8))) +void vsuxseg3ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8mf8_m))) +void vsuxseg3ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8m1))) +void vsuxseg3ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8m1_m))) +void vsuxseg3ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8mf2))) +void vsuxseg3ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8mf2_m))) +void vsuxseg3ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8mf4))) +void vsuxseg3ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8mf4_m))) +void vsuxseg3ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8mf8))) +void vsuxseg3ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8mf8_m))) +void vsuxseg3ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8m1))) +void vsuxseg4ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8m1_m))) +void vsuxseg4ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8mf2))) +void vsuxseg4ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8mf2_m))) +void vsuxseg4ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8mf4))) +void vsuxseg4ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8mf4_m))) +void vsuxseg4ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8mf8))) +void vsuxseg4ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8mf8_m))) +void vsuxseg4ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8m1))) +void vsuxseg4ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8m1_m))) +void vsuxseg4ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8mf2))) +void vsuxseg4ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8mf2_m))) +void vsuxseg4ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8mf4))) +void vsuxseg4ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8mf4_m))) +void vsuxseg4ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8mf8))) +void vsuxseg4ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8mf8_m))) +void vsuxseg4ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8m1))) +void vsuxseg5ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8m1_m))) +void vsuxseg5ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8mf2))) +void vsuxseg5ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8mf2_m))) +void vsuxseg5ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8mf4))) +void vsuxseg5ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8mf4_m))) +void vsuxseg5ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8mf8))) +void vsuxseg5ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8mf8_m))) +void vsuxseg5ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8m1))) +void vsuxseg5ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8m1_m))) +void vsuxseg5ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8mf2))) +void vsuxseg5ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8mf2_m))) +void vsuxseg5ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8mf4))) +void vsuxseg5ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8mf4_m))) +void vsuxseg5ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8mf8))) +void vsuxseg5ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8mf8_m))) +void vsuxseg5ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8m1))) +void vsuxseg6ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8m1_m))) +void vsuxseg6ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8mf2))) +void vsuxseg6ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8mf2_m))) +void vsuxseg6ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8mf4))) +void vsuxseg6ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8mf4_m))) +void vsuxseg6ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8mf8))) +void vsuxseg6ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8mf8_m))) +void vsuxseg6ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8m1))) +void vsuxseg6ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8m1_m))) +void vsuxseg6ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8mf2))) +void vsuxseg6ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8mf2_m))) +void vsuxseg6ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8mf4))) +void vsuxseg6ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8mf4_m))) +void vsuxseg6ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8mf8))) +void vsuxseg6ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8mf8_m))) +void vsuxseg6ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8m1))) +void vsuxseg7ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8m1_m))) +void vsuxseg7ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8mf2))) +void vsuxseg7ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8mf2_m))) +void vsuxseg7ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8mf4))) +void vsuxseg7ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8mf4_m))) +void vsuxseg7ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8mf8))) +void vsuxseg7ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8mf8_m))) +void vsuxseg7ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8m1))) +void vsuxseg7ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8m1_m))) +void vsuxseg7ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8mf2))) +void vsuxseg7ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8mf2_m))) +void vsuxseg7ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8mf4))) +void vsuxseg7ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8mf4_m))) +void vsuxseg7ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8mf8))) +void vsuxseg7ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8mf8_m))) +void vsuxseg7ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8m1))) +void vsuxseg8ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8m1_m))) +void vsuxseg8ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8mf2))) +void vsuxseg8ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8mf2_m))) +void vsuxseg8ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8mf4))) +void vsuxseg8ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8mf4_m))) +void vsuxseg8ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8mf8))) +void vsuxseg8ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8mf8_m))) +void vsuxseg8ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8m1))) +void vsuxseg8ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8m1_m))) +void vsuxseg8ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8mf2))) +void vsuxseg8ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8mf2_m))) +void vsuxseg8ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8mf4))) +void vsuxseg8ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8mf4_m))) +void vsuxseg8ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8mf8))) +void vsuxseg8ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8mf8_m))) +void vsuxseg8ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16m1))) +void vsuxseg2ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16m1_m))) +void vsuxseg2ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16m2))) +void vsuxseg2ei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16m2_m))) +void vsuxseg2ei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16m4))) +void vsuxseg2ei8(int16_t * op0, vuint8m2_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16m4_m))) +void vsuxseg2ei8(vbool4_t op0, int16_t * op1, vuint8m2_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16mf2))) +void vsuxseg2ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16mf2_m))) +void vsuxseg2ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16mf4))) +void vsuxseg2ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16mf4_m))) +void vsuxseg2ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16m1))) +void vsuxseg2ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16m1_m))) +void vsuxseg2ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16m2))) +void vsuxseg2ei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16m2_m))) +void vsuxseg2ei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16m4))) +void vsuxseg2ei8(uint16_t * op0, vuint8m2_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16m4_m))) +void vsuxseg2ei8(vbool4_t op0, uint16_t * op1, vuint8m2_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16mf2))) +void vsuxseg2ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16mf2_m))) +void vsuxseg2ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16mf4))) +void vsuxseg2ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16mf4_m))) +void vsuxseg2ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16m1))) +void vsuxseg3ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16m1_m))) +void vsuxseg3ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16m2))) +void vsuxseg3ei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16m2_m))) +void vsuxseg3ei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16mf2))) +void vsuxseg3ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16mf2_m))) +void vsuxseg3ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16mf4))) +void vsuxseg3ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16mf4_m))) +void vsuxseg3ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16m1))) +void vsuxseg3ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16m1_m))) +void vsuxseg3ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16m2))) +void vsuxseg3ei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16m2_m))) +void vsuxseg3ei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16mf2))) +void vsuxseg3ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16mf2_m))) +void vsuxseg3ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16mf4))) +void vsuxseg3ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16mf4_m))) +void vsuxseg3ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16m1))) +void vsuxseg4ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16m1_m))) +void vsuxseg4ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16m2))) +void vsuxseg4ei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16m2_m))) +void vsuxseg4ei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16mf2))) +void vsuxseg4ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16mf2_m))) +void vsuxseg4ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16mf4))) +void vsuxseg4ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16mf4_m))) +void vsuxseg4ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16m1))) +void vsuxseg4ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16m1_m))) +void vsuxseg4ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16m2))) +void vsuxseg4ei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16m2_m))) +void vsuxseg4ei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16mf2))) +void vsuxseg4ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16mf2_m))) +void vsuxseg4ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16mf4))) +void vsuxseg4ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16mf4_m))) +void vsuxseg4ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i16m1))) +void vsuxseg5ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i16m1_m))) +void vsuxseg5ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i16mf2))) +void vsuxseg5ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i16mf2_m))) +void vsuxseg5ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i16mf4))) +void vsuxseg5ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i16mf4_m))) +void vsuxseg5ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u16m1))) +void vsuxseg5ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u16m1_m))) +void vsuxseg5ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u16mf2))) +void vsuxseg5ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u16mf2_m))) +void vsuxseg5ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u16mf4))) +void vsuxseg5ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u16mf4_m))) +void vsuxseg5ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i16m1))) +void vsuxseg6ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i16m1_m))) +void vsuxseg6ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i16mf2))) +void vsuxseg6ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i16mf2_m))) +void vsuxseg6ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i16mf4))) +void vsuxseg6ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i16mf4_m))) +void vsuxseg6ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u16m1))) +void vsuxseg6ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u16m1_m))) +void vsuxseg6ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u16mf2))) +void vsuxseg6ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u16mf2_m))) +void vsuxseg6ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u16mf4))) +void vsuxseg6ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u16mf4_m))) +void vsuxseg6ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i16m1))) +void vsuxseg7ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i16m1_m))) +void vsuxseg7ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i16mf2))) +void vsuxseg7ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i16mf2_m))) +void vsuxseg7ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i16mf4))) +void vsuxseg7ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i16mf4_m))) +void vsuxseg7ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u16m1))) +void vsuxseg7ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u16m1_m))) +void vsuxseg7ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u16mf2))) +void vsuxseg7ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u16mf2_m))) +void vsuxseg7ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u16mf4))) +void vsuxseg7ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u16mf4_m))) +void vsuxseg7ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i16m1))) +void vsuxseg8ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i16m1_m))) +void vsuxseg8ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i16mf2))) +void vsuxseg8ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i16mf2_m))) +void vsuxseg8ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i16mf4))) +void vsuxseg8ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i16mf4_m))) +void vsuxseg8ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u16m1))) +void vsuxseg8ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u16m1_m))) +void vsuxseg8ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u16mf2))) +void vsuxseg8ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u16mf2_m))) +void vsuxseg8ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u16mf4))) +void vsuxseg8ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u16mf4_m))) +void vsuxseg8ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16m1))) +void vsuxseg2ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16m1_m))) +void vsuxseg2ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16m2))) +void vsuxseg2ei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16m2_m))) +void vsuxseg2ei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16m4))) +void vsuxseg2ei16(int16_t * op0, vuint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16m4_m))) +void vsuxseg2ei16(vbool4_t op0, int16_t * op1, vuint16m4_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16mf2))) +void vsuxseg2ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16mf2_m))) +void vsuxseg2ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16mf4))) +void vsuxseg2ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16mf4_m))) +void vsuxseg2ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16m1))) +void vsuxseg2ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16m1_m))) +void vsuxseg2ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16m2))) +void vsuxseg2ei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16m2_m))) +void vsuxseg2ei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16m4))) +void vsuxseg2ei16(uint16_t * op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16m4_m))) +void vsuxseg2ei16(vbool4_t op0, uint16_t * op1, vuint16m4_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16mf2))) +void vsuxseg2ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16mf2_m))) +void vsuxseg2ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16mf4))) +void vsuxseg2ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16mf4_m))) +void vsuxseg2ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16m1))) +void vsuxseg3ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16m1_m))) +void vsuxseg3ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16m2))) +void vsuxseg3ei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16m2_m))) +void vsuxseg3ei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16mf2))) +void vsuxseg3ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16mf2_m))) +void vsuxseg3ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16mf4))) +void vsuxseg3ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16mf4_m))) +void vsuxseg3ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16m1))) +void vsuxseg3ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16m1_m))) +void vsuxseg3ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16m2))) +void vsuxseg3ei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16m2_m))) +void vsuxseg3ei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16mf2))) +void vsuxseg3ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16mf2_m))) +void vsuxseg3ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16mf4))) +void vsuxseg3ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16mf4_m))) +void vsuxseg3ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16m1))) +void vsuxseg4ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16m1_m))) +void vsuxseg4ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16m2))) +void vsuxseg4ei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16m2_m))) +void vsuxseg4ei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16mf2))) +void vsuxseg4ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16mf2_m))) +void vsuxseg4ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16mf4))) +void vsuxseg4ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16mf4_m))) +void vsuxseg4ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16m1))) +void vsuxseg4ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16m1_m))) +void vsuxseg4ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16m2))) +void vsuxseg4ei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16m2_m))) +void vsuxseg4ei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16mf2))) +void vsuxseg4ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16mf2_m))) +void vsuxseg4ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16mf4))) +void vsuxseg4ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16mf4_m))) +void vsuxseg4ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i16m1))) +void vsuxseg5ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i16m1_m))) +void vsuxseg5ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i16mf2))) +void vsuxseg5ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i16mf2_m))) +void vsuxseg5ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i16mf4))) +void vsuxseg5ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i16mf4_m))) +void vsuxseg5ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u16m1))) +void vsuxseg5ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u16m1_m))) +void vsuxseg5ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u16mf2))) +void vsuxseg5ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u16mf2_m))) +void vsuxseg5ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u16mf4))) +void vsuxseg5ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u16mf4_m))) +void vsuxseg5ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i16m1))) +void vsuxseg6ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i16m1_m))) +void vsuxseg6ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i16mf2))) +void vsuxseg6ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i16mf2_m))) +void vsuxseg6ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i16mf4))) +void vsuxseg6ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i16mf4_m))) +void vsuxseg6ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u16m1))) +void vsuxseg6ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u16m1_m))) +void vsuxseg6ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u16mf2))) +void vsuxseg6ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u16mf2_m))) +void vsuxseg6ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u16mf4))) +void vsuxseg6ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u16mf4_m))) +void vsuxseg6ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i16m1))) +void vsuxseg7ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i16m1_m))) +void vsuxseg7ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i16mf2))) +void vsuxseg7ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i16mf2_m))) +void vsuxseg7ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i16mf4))) +void vsuxseg7ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i16mf4_m))) +void vsuxseg7ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u16m1))) +void vsuxseg7ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u16m1_m))) +void vsuxseg7ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u16mf2))) +void vsuxseg7ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u16mf2_m))) +void vsuxseg7ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u16mf4))) +void vsuxseg7ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u16mf4_m))) +void vsuxseg7ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i16m1))) +void vsuxseg8ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i16m1_m))) +void vsuxseg8ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i16mf2))) +void vsuxseg8ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i16mf2_m))) +void vsuxseg8ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i16mf4))) +void vsuxseg8ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i16mf4_m))) +void vsuxseg8ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u16m1))) +void vsuxseg8ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u16m1_m))) +void vsuxseg8ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u16mf2))) +void vsuxseg8ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u16mf2_m))) +void vsuxseg8ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u16mf4))) +void vsuxseg8ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u16mf4_m))) +void vsuxseg8ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16m1))) +void vsuxseg2ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16m1_m))) +void vsuxseg2ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16m2))) +void vsuxseg2ei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16m2_m))) +void vsuxseg2ei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16m4))) +void vsuxseg2ei32(int16_t * op0, vuint32m8_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16m4_m))) +void vsuxseg2ei32(vbool4_t op0, int16_t * op1, vuint32m8_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16mf2))) +void vsuxseg2ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16mf2_m))) +void vsuxseg2ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16mf4))) +void vsuxseg2ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16mf4_m))) +void vsuxseg2ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16m1))) +void vsuxseg2ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16m1_m))) +void vsuxseg2ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16m2))) +void vsuxseg2ei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16m2_m))) +void vsuxseg2ei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16m4))) +void vsuxseg2ei32(uint16_t * op0, vuint32m8_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16m4_m))) +void vsuxseg2ei32(vbool4_t op0, uint16_t * op1, vuint32m8_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16mf2))) +void vsuxseg2ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16mf2_m))) +void vsuxseg2ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16mf4))) +void vsuxseg2ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16mf4_m))) +void vsuxseg2ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16m1))) +void vsuxseg3ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16m1_m))) +void vsuxseg3ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16m2))) +void vsuxseg3ei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16m2_m))) +void vsuxseg3ei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16mf2))) +void vsuxseg3ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16mf2_m))) +void vsuxseg3ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16mf4))) +void vsuxseg3ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16mf4_m))) +void vsuxseg3ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16m1))) +void vsuxseg3ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16m1_m))) +void vsuxseg3ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16m2))) +void vsuxseg3ei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16m2_m))) +void vsuxseg3ei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16mf2))) +void vsuxseg3ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16mf2_m))) +void vsuxseg3ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16mf4))) +void vsuxseg3ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16mf4_m))) +void vsuxseg3ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16m1))) +void vsuxseg4ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16m1_m))) +void vsuxseg4ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16m2))) +void vsuxseg4ei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16m2_m))) +void vsuxseg4ei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16mf2))) +void vsuxseg4ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16mf2_m))) +void vsuxseg4ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16mf4))) +void vsuxseg4ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16mf4_m))) +void vsuxseg4ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16m1))) +void vsuxseg4ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16m1_m))) +void vsuxseg4ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16m2))) +void vsuxseg4ei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16m2_m))) +void vsuxseg4ei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16mf2))) +void vsuxseg4ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16mf2_m))) +void vsuxseg4ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16mf4))) +void vsuxseg4ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16mf4_m))) +void vsuxseg4ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i16m1))) +void vsuxseg5ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i16m1_m))) +void vsuxseg5ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i16mf2))) +void vsuxseg5ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i16mf2_m))) +void vsuxseg5ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i16mf4))) +void vsuxseg5ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i16mf4_m))) +void vsuxseg5ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u16m1))) +void vsuxseg5ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u16m1_m))) +void vsuxseg5ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u16mf2))) +void vsuxseg5ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u16mf2_m))) +void vsuxseg5ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u16mf4))) +void vsuxseg5ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u16mf4_m))) +void vsuxseg5ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i16m1))) +void vsuxseg6ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i16m1_m))) +void vsuxseg6ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i16mf2))) +void vsuxseg6ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i16mf2_m))) +void vsuxseg6ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i16mf4))) +void vsuxseg6ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i16mf4_m))) +void vsuxseg6ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u16m1))) +void vsuxseg6ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u16m1_m))) +void vsuxseg6ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u16mf2))) +void vsuxseg6ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u16mf2_m))) +void vsuxseg6ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u16mf4))) +void vsuxseg6ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u16mf4_m))) +void vsuxseg6ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i16m1))) +void vsuxseg7ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i16m1_m))) +void vsuxseg7ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i16mf2))) +void vsuxseg7ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i16mf2_m))) +void vsuxseg7ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i16mf4))) +void vsuxseg7ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i16mf4_m))) +void vsuxseg7ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u16m1))) +void vsuxseg7ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u16m1_m))) +void vsuxseg7ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u16mf2))) +void vsuxseg7ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u16mf2_m))) +void vsuxseg7ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u16mf4))) +void vsuxseg7ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u16mf4_m))) +void vsuxseg7ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i16m1))) +void vsuxseg8ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i16m1_m))) +void vsuxseg8ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i16mf2))) +void vsuxseg8ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i16mf2_m))) +void vsuxseg8ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i16mf4))) +void vsuxseg8ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i16mf4_m))) +void vsuxseg8ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u16m1))) +void vsuxseg8ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u16m1_m))) +void vsuxseg8ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u16mf2))) +void vsuxseg8ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u16mf2_m))) +void vsuxseg8ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u16mf4))) +void vsuxseg8ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u16mf4_m))) +void vsuxseg8ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16m1))) +void vsuxseg2ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16m1_m))) +void vsuxseg2ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16m2))) +void vsuxseg2ei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16m2_m))) +void vsuxseg2ei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16mf2))) +void vsuxseg2ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16mf2_m))) +void vsuxseg2ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16mf4))) +void vsuxseg2ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16mf4_m))) +void vsuxseg2ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16m1))) +void vsuxseg2ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16m1_m))) +void vsuxseg2ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16m2))) +void vsuxseg2ei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16m2_m))) +void vsuxseg2ei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16mf2))) +void vsuxseg2ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16mf2_m))) +void vsuxseg2ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16mf4))) +void vsuxseg2ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16mf4_m))) +void vsuxseg2ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16m1))) +void vsuxseg3ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16m1_m))) +void vsuxseg3ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16m2))) +void vsuxseg3ei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16m2_m))) +void vsuxseg3ei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16mf2))) +void vsuxseg3ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16mf2_m))) +void vsuxseg3ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16mf4))) +void vsuxseg3ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16mf4_m))) +void vsuxseg3ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16m1))) +void vsuxseg3ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16m1_m))) +void vsuxseg3ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16m2))) +void vsuxseg3ei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16m2_m))) +void vsuxseg3ei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16mf2))) +void vsuxseg3ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16mf2_m))) +void vsuxseg3ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16mf4))) +void vsuxseg3ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16mf4_m))) +void vsuxseg3ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16m1))) +void vsuxseg4ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16m1_m))) +void vsuxseg4ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16m2))) +void vsuxseg4ei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16m2_m))) +void vsuxseg4ei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16mf2))) +void vsuxseg4ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16mf2_m))) +void vsuxseg4ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16mf4))) +void vsuxseg4ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16mf4_m))) +void vsuxseg4ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16m1))) +void vsuxseg4ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16m1_m))) +void vsuxseg4ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16m2))) +void vsuxseg4ei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16m2_m))) +void vsuxseg4ei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16mf2))) +void vsuxseg4ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16mf2_m))) +void vsuxseg4ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16mf4))) +void vsuxseg4ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16mf4_m))) +void vsuxseg4ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i16m1))) +void vsuxseg5ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i16m1_m))) +void vsuxseg5ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i16mf2))) +void vsuxseg5ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i16mf2_m))) +void vsuxseg5ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i16mf4))) +void vsuxseg5ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i16mf4_m))) +void vsuxseg5ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u16m1))) +void vsuxseg5ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u16m1_m))) +void vsuxseg5ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u16mf2))) +void vsuxseg5ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u16mf2_m))) +void vsuxseg5ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u16mf4))) +void vsuxseg5ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u16mf4_m))) +void vsuxseg5ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i16m1))) +void vsuxseg6ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i16m1_m))) +void vsuxseg6ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i16mf2))) +void vsuxseg6ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i16mf2_m))) +void vsuxseg6ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i16mf4))) +void vsuxseg6ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i16mf4_m))) +void vsuxseg6ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u16m1))) +void vsuxseg6ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u16m1_m))) +void vsuxseg6ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u16mf2))) +void vsuxseg6ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u16mf2_m))) +void vsuxseg6ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u16mf4))) +void vsuxseg6ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u16mf4_m))) +void vsuxseg6ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i16m1))) +void vsuxseg7ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i16m1_m))) +void vsuxseg7ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i16mf2))) +void vsuxseg7ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i16mf2_m))) +void vsuxseg7ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i16mf4))) +void vsuxseg7ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i16mf4_m))) +void vsuxseg7ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u16m1))) +void vsuxseg7ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u16m1_m))) +void vsuxseg7ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u16mf2))) +void vsuxseg7ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u16mf2_m))) +void vsuxseg7ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u16mf4))) +void vsuxseg7ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u16mf4_m))) +void vsuxseg7ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i16m1))) +void vsuxseg8ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i16m1_m))) +void vsuxseg8ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i16mf2))) +void vsuxseg8ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i16mf2_m))) +void vsuxseg8ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i16mf4))) +void vsuxseg8ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i16mf4_m))) +void vsuxseg8ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u16m1))) +void vsuxseg8ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u16m1_m))) +void vsuxseg8ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u16mf2))) +void vsuxseg8ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u16mf2_m))) +void vsuxseg8ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u16mf4))) +void vsuxseg8ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u16mf4_m))) +void vsuxseg8ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32m1))) +void vsuxseg2ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32m1_m))) +void vsuxseg2ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32m2))) +void vsuxseg2ei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32m2_m))) +void vsuxseg2ei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32m4))) +void vsuxseg2ei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32m4_m))) +void vsuxseg2ei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32mf2))) +void vsuxseg2ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32mf2_m))) +void vsuxseg2ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32m1))) +void vsuxseg2ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32m1_m))) +void vsuxseg2ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32m2))) +void vsuxseg2ei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32m2_m))) +void vsuxseg2ei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32m4))) +void vsuxseg2ei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32m4_m))) +void vsuxseg2ei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32mf2))) +void vsuxseg2ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32mf2_m))) +void vsuxseg2ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i32m1))) +void vsuxseg3ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i32m1_m))) +void vsuxseg3ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i32m2))) +void vsuxseg3ei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i32m2_m))) +void vsuxseg3ei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i32mf2))) +void vsuxseg3ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i32mf2_m))) +void vsuxseg3ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u32m1))) +void vsuxseg3ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u32m1_m))) +void vsuxseg3ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u32m2))) +void vsuxseg3ei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u32m2_m))) +void vsuxseg3ei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u32mf2))) +void vsuxseg3ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u32mf2_m))) +void vsuxseg3ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i32m1))) +void vsuxseg4ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i32m1_m))) +void vsuxseg4ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i32m2))) +void vsuxseg4ei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i32m2_m))) +void vsuxseg4ei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i32mf2))) +void vsuxseg4ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i32mf2_m))) +void vsuxseg4ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u32m1))) +void vsuxseg4ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u32m1_m))) +void vsuxseg4ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u32m2))) +void vsuxseg4ei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u32m2_m))) +void vsuxseg4ei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u32mf2))) +void vsuxseg4ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u32mf2_m))) +void vsuxseg4ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i32m1))) +void vsuxseg5ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i32m1_m))) +void vsuxseg5ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i32mf2))) +void vsuxseg5ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i32mf2_m))) +void vsuxseg5ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u32m1))) +void vsuxseg5ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u32m1_m))) +void vsuxseg5ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u32mf2))) +void vsuxseg5ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u32mf2_m))) +void vsuxseg5ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i32m1))) +void vsuxseg6ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i32m1_m))) +void vsuxseg6ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i32mf2))) +void vsuxseg6ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i32mf2_m))) +void vsuxseg6ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u32m1))) +void vsuxseg6ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u32m1_m))) +void vsuxseg6ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u32mf2))) +void vsuxseg6ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u32mf2_m))) +void vsuxseg6ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i32m1))) +void vsuxseg7ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i32m1_m))) +void vsuxseg7ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i32mf2))) +void vsuxseg7ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i32mf2_m))) +void vsuxseg7ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u32m1))) +void vsuxseg7ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u32m1_m))) +void vsuxseg7ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u32mf2))) +void vsuxseg7ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u32mf2_m))) +void vsuxseg7ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i32m1))) +void vsuxseg8ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i32m1_m))) +void vsuxseg8ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i32mf2))) +void vsuxseg8ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i32mf2_m))) +void vsuxseg8ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u32m1))) +void vsuxseg8ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u32m1_m))) +void vsuxseg8ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u32mf2))) +void vsuxseg8ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u32mf2_m))) +void vsuxseg8ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32m1))) +void vsuxseg2ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32m1_m))) +void vsuxseg2ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32m2))) +void vsuxseg2ei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32m2_m))) +void vsuxseg2ei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32m4))) +void vsuxseg2ei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32m4_m))) +void vsuxseg2ei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32mf2))) +void vsuxseg2ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32mf2_m))) +void vsuxseg2ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32m1))) +void vsuxseg2ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32m1_m))) +void vsuxseg2ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32m2))) +void vsuxseg2ei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32m2_m))) +void vsuxseg2ei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32m4))) +void vsuxseg2ei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32m4_m))) +void vsuxseg2ei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32mf2))) +void vsuxseg2ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32mf2_m))) +void vsuxseg2ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i32m1))) +void vsuxseg3ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i32m1_m))) +void vsuxseg3ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i32m2))) +void vsuxseg3ei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i32m2_m))) +void vsuxseg3ei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i32mf2))) +void vsuxseg3ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i32mf2_m))) +void vsuxseg3ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u32m1))) +void vsuxseg3ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u32m1_m))) +void vsuxseg3ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u32m2))) +void vsuxseg3ei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u32m2_m))) +void vsuxseg3ei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u32mf2))) +void vsuxseg3ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u32mf2_m))) +void vsuxseg3ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i32m1))) +void vsuxseg4ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i32m1_m))) +void vsuxseg4ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i32m2))) +void vsuxseg4ei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i32m2_m))) +void vsuxseg4ei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i32mf2))) +void vsuxseg4ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i32mf2_m))) +void vsuxseg4ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u32m1))) +void vsuxseg4ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u32m1_m))) +void vsuxseg4ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u32m2))) +void vsuxseg4ei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u32m2_m))) +void vsuxseg4ei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u32mf2))) +void vsuxseg4ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u32mf2_m))) +void vsuxseg4ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i32m1))) +void vsuxseg5ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i32m1_m))) +void vsuxseg5ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i32mf2))) +void vsuxseg5ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i32mf2_m))) +void vsuxseg5ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u32m1))) +void vsuxseg5ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u32m1_m))) +void vsuxseg5ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u32mf2))) +void vsuxseg5ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u32mf2_m))) +void vsuxseg5ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i32m1))) +void vsuxseg6ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i32m1_m))) +void vsuxseg6ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i32mf2))) +void vsuxseg6ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i32mf2_m))) +void vsuxseg6ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u32m1))) +void vsuxseg6ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u32m1_m))) +void vsuxseg6ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u32mf2))) +void vsuxseg6ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u32mf2_m))) +void vsuxseg6ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i32m1))) +void vsuxseg7ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i32m1_m))) +void vsuxseg7ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i32mf2))) +void vsuxseg7ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i32mf2_m))) +void vsuxseg7ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u32m1))) +void vsuxseg7ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u32m1_m))) +void vsuxseg7ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u32mf2))) +void vsuxseg7ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u32mf2_m))) +void vsuxseg7ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i32m1))) +void vsuxseg8ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i32m1_m))) +void vsuxseg8ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i32mf2))) +void vsuxseg8ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i32mf2_m))) +void vsuxseg8ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u32m1))) +void vsuxseg8ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u32m1_m))) +void vsuxseg8ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u32mf2))) +void vsuxseg8ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u32mf2_m))) +void vsuxseg8ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32m1))) +void vsuxseg2ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32m1_m))) +void vsuxseg2ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32m2))) +void vsuxseg2ei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32m2_m))) +void vsuxseg2ei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32m4))) +void vsuxseg2ei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32m4_m))) +void vsuxseg2ei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32mf2))) +void vsuxseg2ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32mf2_m))) +void vsuxseg2ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32m1))) +void vsuxseg2ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32m1_m))) +void vsuxseg2ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32m2))) +void vsuxseg2ei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32m2_m))) +void vsuxseg2ei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32m4))) +void vsuxseg2ei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32m4_m))) +void vsuxseg2ei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32mf2))) +void vsuxseg2ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32mf2_m))) +void vsuxseg2ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i32m1))) +void vsuxseg3ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i32m1_m))) +void vsuxseg3ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i32m2))) +void vsuxseg3ei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i32m2_m))) +void vsuxseg3ei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i32mf2))) +void vsuxseg3ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i32mf2_m))) +void vsuxseg3ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u32m1))) +void vsuxseg3ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u32m1_m))) +void vsuxseg3ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u32m2))) +void vsuxseg3ei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u32m2_m))) +void vsuxseg3ei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u32mf2))) +void vsuxseg3ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u32mf2_m))) +void vsuxseg3ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i32m1))) +void vsuxseg4ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i32m1_m))) +void vsuxseg4ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i32m2))) +void vsuxseg4ei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i32m2_m))) +void vsuxseg4ei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i32mf2))) +void vsuxseg4ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i32mf2_m))) +void vsuxseg4ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u32m1))) +void vsuxseg4ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u32m1_m))) +void vsuxseg4ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u32m2))) +void vsuxseg4ei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u32m2_m))) +void vsuxseg4ei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u32mf2))) +void vsuxseg4ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u32mf2_m))) +void vsuxseg4ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i32m1))) +void vsuxseg5ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i32m1_m))) +void vsuxseg5ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i32mf2))) +void vsuxseg5ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i32mf2_m))) +void vsuxseg5ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u32m1))) +void vsuxseg5ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u32m1_m))) +void vsuxseg5ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u32mf2))) +void vsuxseg5ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u32mf2_m))) +void vsuxseg5ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i32m1))) +void vsuxseg6ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i32m1_m))) +void vsuxseg6ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i32mf2))) +void vsuxseg6ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i32mf2_m))) +void vsuxseg6ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u32m1))) +void vsuxseg6ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u32m1_m))) +void vsuxseg6ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u32mf2))) +void vsuxseg6ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u32mf2_m))) +void vsuxseg6ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i32m1))) +void vsuxseg7ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i32m1_m))) +void vsuxseg7ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i32mf2))) +void vsuxseg7ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i32mf2_m))) +void vsuxseg7ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u32m1))) +void vsuxseg7ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u32m1_m))) +void vsuxseg7ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u32mf2))) +void vsuxseg7ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u32mf2_m))) +void vsuxseg7ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i32m1))) +void vsuxseg8ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i32m1_m))) +void vsuxseg8ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i32mf2))) +void vsuxseg8ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i32mf2_m))) +void vsuxseg8ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u32m1))) +void vsuxseg8ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u32m1_m))) +void vsuxseg8ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u32mf2))) +void vsuxseg8ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u32mf2_m))) +void vsuxseg8ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32m1))) +void vsuxseg2ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32m1_m))) +void vsuxseg2ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32m2))) +void vsuxseg2ei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32m2_m))) +void vsuxseg2ei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32m4))) +void vsuxseg2ei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32m4_m))) +void vsuxseg2ei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32mf2))) +void vsuxseg2ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32mf2_m))) +void vsuxseg2ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32m1))) +void vsuxseg2ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32m1_m))) +void vsuxseg2ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32m2))) +void vsuxseg2ei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32m2_m))) +void vsuxseg2ei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32m4))) +void vsuxseg2ei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32m4_m))) +void vsuxseg2ei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32mf2))) +void vsuxseg2ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32mf2_m))) +void vsuxseg2ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i32m1))) +void vsuxseg3ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i32m1_m))) +void vsuxseg3ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i32m2))) +void vsuxseg3ei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i32m2_m))) +void vsuxseg3ei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i32mf2))) +void vsuxseg3ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i32mf2_m))) +void vsuxseg3ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u32m1))) +void vsuxseg3ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u32m1_m))) +void vsuxseg3ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u32m2))) +void vsuxseg3ei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u32m2_m))) +void vsuxseg3ei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u32mf2))) +void vsuxseg3ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u32mf2_m))) +void vsuxseg3ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i32m1))) +void vsuxseg4ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i32m1_m))) +void vsuxseg4ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i32m2))) +void vsuxseg4ei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i32m2_m))) +void vsuxseg4ei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i32mf2))) +void vsuxseg4ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i32mf2_m))) +void vsuxseg4ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u32m1))) +void vsuxseg4ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u32m1_m))) +void vsuxseg4ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u32m2))) +void vsuxseg4ei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u32m2_m))) +void vsuxseg4ei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u32mf2))) +void vsuxseg4ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u32mf2_m))) +void vsuxseg4ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i32m1))) +void vsuxseg5ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i32m1_m))) +void vsuxseg5ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i32mf2))) +void vsuxseg5ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i32mf2_m))) +void vsuxseg5ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u32m1))) +void vsuxseg5ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u32m1_m))) +void vsuxseg5ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u32mf2))) +void vsuxseg5ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u32mf2_m))) +void vsuxseg5ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i32m1))) +void vsuxseg6ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i32m1_m))) +void vsuxseg6ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i32mf2))) +void vsuxseg6ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i32mf2_m))) +void vsuxseg6ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u32m1))) +void vsuxseg6ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u32m1_m))) +void vsuxseg6ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u32mf2))) +void vsuxseg6ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u32mf2_m))) +void vsuxseg6ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i32m1))) +void vsuxseg7ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i32m1_m))) +void vsuxseg7ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i32mf2))) +void vsuxseg7ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i32mf2_m))) +void vsuxseg7ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u32m1))) +void vsuxseg7ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u32m1_m))) +void vsuxseg7ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u32mf2))) +void vsuxseg7ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u32mf2_m))) +void vsuxseg7ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i32m1))) +void vsuxseg8ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i32m1_m))) +void vsuxseg8ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i32mf2))) +void vsuxseg8ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i32mf2_m))) +void vsuxseg8ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u32m1))) +void vsuxseg8ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u32m1_m))) +void vsuxseg8ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u32mf2))) +void vsuxseg8ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u32mf2_m))) +void vsuxseg8ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i64m1))) +void vsuxseg2ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i64m1_m))) +void vsuxseg2ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i64m2))) +void vsuxseg2ei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i64m2_m))) +void vsuxseg2ei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i64m4))) +void vsuxseg2ei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i64m4_m))) +void vsuxseg2ei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u64m1))) +void vsuxseg2ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u64m1_m))) +void vsuxseg2ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u64m2))) +void vsuxseg2ei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u64m2_m))) +void vsuxseg2ei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u64m4))) +void vsuxseg2ei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u64m4_m))) +void vsuxseg2ei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i64m1))) +void vsuxseg3ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i64m1_m))) +void vsuxseg3ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i64m2))) +void vsuxseg3ei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i64m2_m))) +void vsuxseg3ei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u64m1))) +void vsuxseg3ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u64m1_m))) +void vsuxseg3ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u64m2))) +void vsuxseg3ei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u64m2_m))) +void vsuxseg3ei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i64m1))) +void vsuxseg4ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i64m1_m))) +void vsuxseg4ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i64m2))) +void vsuxseg4ei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i64m2_m))) +void vsuxseg4ei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u64m1))) +void vsuxseg4ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u64m1_m))) +void vsuxseg4ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u64m2))) +void vsuxseg4ei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u64m2_m))) +void vsuxseg4ei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i64m1))) +void vsuxseg5ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i64m1_m))) +void vsuxseg5ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u64m1))) +void vsuxseg5ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u64m1_m))) +void vsuxseg5ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i64m1))) +void vsuxseg6ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i64m1_m))) +void vsuxseg6ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u64m1))) +void vsuxseg6ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u64m1_m))) +void vsuxseg6ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i64m1))) +void vsuxseg7ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i64m1_m))) +void vsuxseg7ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u64m1))) +void vsuxseg7ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u64m1_m))) +void vsuxseg7ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i64m1))) +void vsuxseg8ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i64m1_m))) +void vsuxseg8ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u64m1))) +void vsuxseg8ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u64m1_m))) +void vsuxseg8ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i64m1))) +void vsuxseg2ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i64m1_m))) +void vsuxseg2ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i64m2))) +void vsuxseg2ei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i64m2_m))) +void vsuxseg2ei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i64m4))) +void vsuxseg2ei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i64m4_m))) +void vsuxseg2ei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u64m1))) +void vsuxseg2ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u64m1_m))) +void vsuxseg2ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u64m2))) +void vsuxseg2ei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u64m2_m))) +void vsuxseg2ei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u64m4))) +void vsuxseg2ei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u64m4_m))) +void vsuxseg2ei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i64m1))) +void vsuxseg3ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i64m1_m))) +void vsuxseg3ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i64m2))) +void vsuxseg3ei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i64m2_m))) +void vsuxseg3ei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u64m1))) +void vsuxseg3ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u64m1_m))) +void vsuxseg3ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u64m2))) +void vsuxseg3ei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u64m2_m))) +void vsuxseg3ei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i64m1))) +void vsuxseg4ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i64m1_m))) +void vsuxseg4ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i64m2))) +void vsuxseg4ei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i64m2_m))) +void vsuxseg4ei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u64m1))) +void vsuxseg4ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u64m1_m))) +void vsuxseg4ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u64m2))) +void vsuxseg4ei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u64m2_m))) +void vsuxseg4ei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i64m1))) +void vsuxseg5ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i64m1_m))) +void vsuxseg5ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u64m1))) +void vsuxseg5ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u64m1_m))) +void vsuxseg5ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i64m1))) +void vsuxseg6ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i64m1_m))) +void vsuxseg6ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u64m1))) +void vsuxseg6ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u64m1_m))) +void vsuxseg6ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i64m1))) +void vsuxseg7ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i64m1_m))) +void vsuxseg7ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u64m1))) +void vsuxseg7ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u64m1_m))) +void vsuxseg7ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i64m1))) +void vsuxseg8ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i64m1_m))) +void vsuxseg8ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u64m1))) +void vsuxseg8ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u64m1_m))) +void vsuxseg8ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i64m1))) +void vsuxseg2ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i64m1_m))) +void vsuxseg2ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i64m2))) +void vsuxseg2ei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i64m2_m))) +void vsuxseg2ei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i64m4))) +void vsuxseg2ei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i64m4_m))) +void vsuxseg2ei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u64m1))) +void vsuxseg2ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u64m1_m))) +void vsuxseg2ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u64m2))) +void vsuxseg2ei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u64m2_m))) +void vsuxseg2ei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u64m4))) +void vsuxseg2ei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u64m4_m))) +void vsuxseg2ei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i64m1))) +void vsuxseg3ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i64m1_m))) +void vsuxseg3ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i64m2))) +void vsuxseg3ei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i64m2_m))) +void vsuxseg3ei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u64m1))) +void vsuxseg3ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u64m1_m))) +void vsuxseg3ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u64m2))) +void vsuxseg3ei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u64m2_m))) +void vsuxseg3ei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i64m1))) +void vsuxseg4ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i64m1_m))) +void vsuxseg4ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i64m2))) +void vsuxseg4ei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i64m2_m))) +void vsuxseg4ei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u64m1))) +void vsuxseg4ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u64m1_m))) +void vsuxseg4ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u64m2))) +void vsuxseg4ei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u64m2_m))) +void vsuxseg4ei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i64m1))) +void vsuxseg5ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i64m1_m))) +void vsuxseg5ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u64m1))) +void vsuxseg5ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u64m1_m))) +void vsuxseg5ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i64m1))) +void vsuxseg6ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i64m1_m))) +void vsuxseg6ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u64m1))) +void vsuxseg6ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u64m1_m))) +void vsuxseg6ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i64m1))) +void vsuxseg7ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i64m1_m))) +void vsuxseg7ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u64m1))) +void vsuxseg7ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u64m1_m))) +void vsuxseg7ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i64m1))) +void vsuxseg8ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i64m1_m))) +void vsuxseg8ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u64m1))) +void vsuxseg8ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u64m1_m))) +void vsuxseg8ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i64m1))) +void vsuxseg2ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i64m1_m))) +void vsuxseg2ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i64m2))) +void vsuxseg2ei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i64m2_m))) +void vsuxseg2ei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i64m4))) +void vsuxseg2ei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i64m4_m))) +void vsuxseg2ei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u64m1))) +void vsuxseg2ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u64m1_m))) +void vsuxseg2ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u64m2))) +void vsuxseg2ei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u64m2_m))) +void vsuxseg2ei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u64m4))) +void vsuxseg2ei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u64m4_m))) +void vsuxseg2ei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i64m1))) +void vsuxseg3ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i64m1_m))) +void vsuxseg3ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i64m2))) +void vsuxseg3ei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i64m2_m))) +void vsuxseg3ei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u64m1))) +void vsuxseg3ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u64m1_m))) +void vsuxseg3ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u64m2))) +void vsuxseg3ei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u64m2_m))) +void vsuxseg3ei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i64m1))) +void vsuxseg4ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i64m1_m))) +void vsuxseg4ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i64m2))) +void vsuxseg4ei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i64m2_m))) +void vsuxseg4ei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u64m1))) +void vsuxseg4ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u64m1_m))) +void vsuxseg4ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u64m2))) +void vsuxseg4ei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u64m2_m))) +void vsuxseg4ei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i64m1))) +void vsuxseg5ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i64m1_m))) +void vsuxseg5ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u64m1))) +void vsuxseg5ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u64m1_m))) +void vsuxseg5ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i64m1))) +void vsuxseg6ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i64m1_m))) +void vsuxseg6ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u64m1))) +void vsuxseg6ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u64m1_m))) +void vsuxseg6ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i64m1))) +void vsuxseg7ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i64m1_m))) +void vsuxseg7ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u64m1))) +void vsuxseg7ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u64m1_m))) +void vsuxseg7ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i64m1))) +void vsuxseg8ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i64m1_m))) +void vsuxseg8ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u64m1))) +void vsuxseg8ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u64m1_m))) +void vsuxseg8ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8m1))) +void vsoxseg2ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8m1_m))) +void vsoxseg2ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8m2))) +void vsoxseg2ei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8m2_m))) +void vsoxseg2ei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8m4))) +void vsoxseg2ei8(int8_t * op0, vuint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8m4_m))) +void vsoxseg2ei8(vbool2_t op0, int8_t * op1, vuint8m4_t op2, vint8m4_t op3, vint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8mf2))) +void vsoxseg2ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8mf2_m))) +void vsoxseg2ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8mf4))) +void vsoxseg2ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8mf4_m))) +void vsoxseg2ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8mf8))) +void vsoxseg2ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8mf8_m))) +void vsoxseg2ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8m1))) +void vsoxseg2ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8m1_m))) +void vsoxseg2ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8m2))) +void vsoxseg2ei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8m2_m))) +void vsoxseg2ei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8m4))) +void vsoxseg2ei8(uint8_t * op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8m4_m))) +void vsoxseg2ei8(vbool2_t op0, uint8_t * op1, vuint8m4_t op2, vuint8m4_t op3, vuint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8mf2))) +void vsoxseg2ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8mf2_m))) +void vsoxseg2ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8mf4))) +void vsoxseg2ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8mf4_m))) +void vsoxseg2ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8mf8))) +void vsoxseg2ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8mf8_m))) +void vsoxseg2ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8m1))) +void vsoxseg3ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8m1_m))) +void vsoxseg3ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8m2))) +void vsoxseg3ei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8m2_m))) +void vsoxseg3ei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8mf2))) +void vsoxseg3ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8mf2_m))) +void vsoxseg3ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8mf4))) +void vsoxseg3ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8mf4_m))) +void vsoxseg3ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8mf8))) +void vsoxseg3ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8mf8_m))) +void vsoxseg3ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8m1))) +void vsoxseg3ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8m1_m))) +void vsoxseg3ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8m2))) +void vsoxseg3ei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8m2_m))) +void vsoxseg3ei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8mf2))) +void vsoxseg3ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8mf2_m))) +void vsoxseg3ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8mf4))) +void vsoxseg3ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8mf4_m))) +void vsoxseg3ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8mf8))) +void vsoxseg3ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8mf8_m))) +void vsoxseg3ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8m1))) +void vsoxseg4ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8m1_m))) +void vsoxseg4ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8m2))) +void vsoxseg4ei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8m2_m))) +void vsoxseg4ei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8mf2))) +void vsoxseg4ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8mf2_m))) +void vsoxseg4ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8mf4))) +void vsoxseg4ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8mf4_m))) +void vsoxseg4ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8mf8))) +void vsoxseg4ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8mf8_m))) +void vsoxseg4ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8m1))) +void vsoxseg4ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8m1_m))) +void vsoxseg4ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8m2))) +void vsoxseg4ei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8m2_m))) +void vsoxseg4ei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8mf2))) +void vsoxseg4ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8mf2_m))) +void vsoxseg4ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8mf4))) +void vsoxseg4ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8mf4_m))) +void vsoxseg4ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8mf8))) +void vsoxseg4ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8mf8_m))) +void vsoxseg4ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8m1))) +void vsoxseg5ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8m1_m))) +void vsoxseg5ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8mf2))) +void vsoxseg5ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8mf2_m))) +void vsoxseg5ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8mf4))) +void vsoxseg5ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8mf4_m))) +void vsoxseg5ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8mf8))) +void vsoxseg5ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8mf8_m))) +void vsoxseg5ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8m1))) +void vsoxseg5ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8m1_m))) +void vsoxseg5ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8mf2))) +void vsoxseg5ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8mf2_m))) +void vsoxseg5ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8mf4))) +void vsoxseg5ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8mf4_m))) +void vsoxseg5ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8mf8))) +void vsoxseg5ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8mf8_m))) +void vsoxseg5ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8m1))) +void vsoxseg6ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8m1_m))) +void vsoxseg6ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8mf2))) +void vsoxseg6ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8mf2_m))) +void vsoxseg6ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8mf4))) +void vsoxseg6ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8mf4_m))) +void vsoxseg6ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8mf8))) +void vsoxseg6ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8mf8_m))) +void vsoxseg6ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8m1))) +void vsoxseg6ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8m1_m))) +void vsoxseg6ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8mf2))) +void vsoxseg6ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8mf2_m))) +void vsoxseg6ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8mf4))) +void vsoxseg6ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8mf4_m))) +void vsoxseg6ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8mf8))) +void vsoxseg6ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8mf8_m))) +void vsoxseg6ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8m1))) +void vsoxseg7ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8m1_m))) +void vsoxseg7ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8mf2))) +void vsoxseg7ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8mf2_m))) +void vsoxseg7ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8mf4))) +void vsoxseg7ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8mf4_m))) +void vsoxseg7ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8mf8))) +void vsoxseg7ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8mf8_m))) +void vsoxseg7ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8m1))) +void vsoxseg7ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8m1_m))) +void vsoxseg7ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8mf2))) +void vsoxseg7ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8mf2_m))) +void vsoxseg7ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8mf4))) +void vsoxseg7ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8mf4_m))) +void vsoxseg7ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8mf8))) +void vsoxseg7ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8mf8_m))) +void vsoxseg7ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8m1))) +void vsoxseg8ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8m1_m))) +void vsoxseg8ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8mf2))) +void vsoxseg8ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8mf2_m))) +void vsoxseg8ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8mf4))) +void vsoxseg8ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8mf4_m))) +void vsoxseg8ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8mf8))) +void vsoxseg8ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8mf8_m))) +void vsoxseg8ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8m1))) +void vsoxseg8ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8m1_m))) +void vsoxseg8ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8mf2))) +void vsoxseg8ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8mf2_m))) +void vsoxseg8ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8mf4))) +void vsoxseg8ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8mf4_m))) +void vsoxseg8ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8mf8))) +void vsoxseg8ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8mf8_m))) +void vsoxseg8ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8m1))) +void vsoxseg2ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8m1_m))) +void vsoxseg2ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8m2))) +void vsoxseg2ei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8m2_m))) +void vsoxseg2ei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8m4))) +void vsoxseg2ei16(int8_t * op0, vuint16m8_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8m4_m))) +void vsoxseg2ei16(vbool2_t op0, int8_t * op1, vuint16m8_t op2, vint8m4_t op3, vint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8mf2))) +void vsoxseg2ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8mf2_m))) +void vsoxseg2ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8mf4))) +void vsoxseg2ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8mf4_m))) +void vsoxseg2ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8mf8))) +void vsoxseg2ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8mf8_m))) +void vsoxseg2ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8m1))) +void vsoxseg2ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8m1_m))) +void vsoxseg2ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8m2))) +void vsoxseg2ei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8m2_m))) +void vsoxseg2ei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8m4))) +void vsoxseg2ei16(uint8_t * op0, vuint16m8_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8m4_m))) +void vsoxseg2ei16(vbool2_t op0, uint8_t * op1, vuint16m8_t op2, vuint8m4_t op3, vuint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8mf2))) +void vsoxseg2ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8mf2_m))) +void vsoxseg2ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8mf4))) +void vsoxseg2ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8mf4_m))) +void vsoxseg2ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8mf8))) +void vsoxseg2ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8mf8_m))) +void vsoxseg2ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8m1))) +void vsoxseg3ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8m1_m))) +void vsoxseg3ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8m2))) +void vsoxseg3ei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8m2_m))) +void vsoxseg3ei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8mf2))) +void vsoxseg3ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8mf2_m))) +void vsoxseg3ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8mf4))) +void vsoxseg3ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8mf4_m))) +void vsoxseg3ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8mf8))) +void vsoxseg3ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8mf8_m))) +void vsoxseg3ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8m1))) +void vsoxseg3ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8m1_m))) +void vsoxseg3ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8m2))) +void vsoxseg3ei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8m2_m))) +void vsoxseg3ei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8mf2))) +void vsoxseg3ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8mf2_m))) +void vsoxseg3ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8mf4))) +void vsoxseg3ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8mf4_m))) +void vsoxseg3ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8mf8))) +void vsoxseg3ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8mf8_m))) +void vsoxseg3ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8m1))) +void vsoxseg4ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8m1_m))) +void vsoxseg4ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8m2))) +void vsoxseg4ei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8m2_m))) +void vsoxseg4ei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8mf2))) +void vsoxseg4ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8mf2_m))) +void vsoxseg4ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8mf4))) +void vsoxseg4ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8mf4_m))) +void vsoxseg4ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8mf8))) +void vsoxseg4ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8mf8_m))) +void vsoxseg4ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8m1))) +void vsoxseg4ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8m1_m))) +void vsoxseg4ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8m2))) +void vsoxseg4ei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8m2_m))) +void vsoxseg4ei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8mf2))) +void vsoxseg4ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8mf2_m))) +void vsoxseg4ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8mf4))) +void vsoxseg4ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8mf4_m))) +void vsoxseg4ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8mf8))) +void vsoxseg4ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8mf8_m))) +void vsoxseg4ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8m1))) +void vsoxseg5ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8m1_m))) +void vsoxseg5ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8mf2))) +void vsoxseg5ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8mf2_m))) +void vsoxseg5ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8mf4))) +void vsoxseg5ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8mf4_m))) +void vsoxseg5ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8mf8))) +void vsoxseg5ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8mf8_m))) +void vsoxseg5ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8m1))) +void vsoxseg5ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8m1_m))) +void vsoxseg5ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8mf2))) +void vsoxseg5ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8mf2_m))) +void vsoxseg5ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8mf4))) +void vsoxseg5ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8mf4_m))) +void vsoxseg5ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8mf8))) +void vsoxseg5ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8mf8_m))) +void vsoxseg5ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8m1))) +void vsoxseg6ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8m1_m))) +void vsoxseg6ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8mf2))) +void vsoxseg6ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8mf2_m))) +void vsoxseg6ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8mf4))) +void vsoxseg6ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8mf4_m))) +void vsoxseg6ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8mf8))) +void vsoxseg6ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8mf8_m))) +void vsoxseg6ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8m1))) +void vsoxseg6ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8m1_m))) +void vsoxseg6ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8mf2))) +void vsoxseg6ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8mf2_m))) +void vsoxseg6ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8mf4))) +void vsoxseg6ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8mf4_m))) +void vsoxseg6ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8mf8))) +void vsoxseg6ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8mf8_m))) +void vsoxseg6ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8m1))) +void vsoxseg7ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8m1_m))) +void vsoxseg7ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8mf2))) +void vsoxseg7ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8mf2_m))) +void vsoxseg7ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8mf4))) +void vsoxseg7ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8mf4_m))) +void vsoxseg7ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8mf8))) +void vsoxseg7ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8mf8_m))) +void vsoxseg7ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8m1))) +void vsoxseg7ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8m1_m))) +void vsoxseg7ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8mf2))) +void vsoxseg7ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8mf2_m))) +void vsoxseg7ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8mf4))) +void vsoxseg7ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8mf4_m))) +void vsoxseg7ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8mf8))) +void vsoxseg7ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8mf8_m))) +void vsoxseg7ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8m1))) +void vsoxseg8ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8m1_m))) +void vsoxseg8ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8mf2))) +void vsoxseg8ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8mf2_m))) +void vsoxseg8ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8mf4))) +void vsoxseg8ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8mf4_m))) +void vsoxseg8ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8mf8))) +void vsoxseg8ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8mf8_m))) +void vsoxseg8ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8m1))) +void vsoxseg8ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8m1_m))) +void vsoxseg8ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8mf2))) +void vsoxseg8ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8mf2_m))) +void vsoxseg8ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8mf4))) +void vsoxseg8ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8mf4_m))) +void vsoxseg8ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8mf8))) +void vsoxseg8ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8mf8_m))) +void vsoxseg8ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8m1))) +void vsoxseg2ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8m1_m))) +void vsoxseg2ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8m2))) +void vsoxseg2ei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8m2_m))) +void vsoxseg2ei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8mf2))) +void vsoxseg2ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8mf2_m))) +void vsoxseg2ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8mf4))) +void vsoxseg2ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8mf4_m))) +void vsoxseg2ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8mf8))) +void vsoxseg2ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8mf8_m))) +void vsoxseg2ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8m1))) +void vsoxseg2ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8m1_m))) +void vsoxseg2ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8m2))) +void vsoxseg2ei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8m2_m))) +void vsoxseg2ei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8mf2))) +void vsoxseg2ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8mf2_m))) +void vsoxseg2ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8mf4))) +void vsoxseg2ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8mf4_m))) +void vsoxseg2ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8mf8))) +void vsoxseg2ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8mf8_m))) +void vsoxseg2ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8m1))) +void vsoxseg3ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8m1_m))) +void vsoxseg3ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8m2))) +void vsoxseg3ei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8m2_m))) +void vsoxseg3ei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8mf2))) +void vsoxseg3ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8mf2_m))) +void vsoxseg3ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8mf4))) +void vsoxseg3ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8mf4_m))) +void vsoxseg3ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8mf8))) +void vsoxseg3ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8mf8_m))) +void vsoxseg3ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8m1))) +void vsoxseg3ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8m1_m))) +void vsoxseg3ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8m2))) +void vsoxseg3ei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8m2_m))) +void vsoxseg3ei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8mf2))) +void vsoxseg3ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8mf2_m))) +void vsoxseg3ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8mf4))) +void vsoxseg3ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8mf4_m))) +void vsoxseg3ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8mf8))) +void vsoxseg3ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8mf8_m))) +void vsoxseg3ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8m1))) +void vsoxseg4ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8m1_m))) +void vsoxseg4ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8m2))) +void vsoxseg4ei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8m2_m))) +void vsoxseg4ei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8mf2))) +void vsoxseg4ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8mf2_m))) +void vsoxseg4ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8mf4))) +void vsoxseg4ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8mf4_m))) +void vsoxseg4ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8mf8))) +void vsoxseg4ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8mf8_m))) +void vsoxseg4ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8m1))) +void vsoxseg4ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8m1_m))) +void vsoxseg4ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8m2))) +void vsoxseg4ei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8m2_m))) +void vsoxseg4ei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8mf2))) +void vsoxseg4ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8mf2_m))) +void vsoxseg4ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8mf4))) +void vsoxseg4ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8mf4_m))) +void vsoxseg4ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8mf8))) +void vsoxseg4ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8mf8_m))) +void vsoxseg4ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8m1))) +void vsoxseg5ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8m1_m))) +void vsoxseg5ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8mf2))) +void vsoxseg5ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8mf2_m))) +void vsoxseg5ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8mf4))) +void vsoxseg5ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8mf4_m))) +void vsoxseg5ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8mf8))) +void vsoxseg5ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8mf8_m))) +void vsoxseg5ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8m1))) +void vsoxseg5ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8m1_m))) +void vsoxseg5ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8mf2))) +void vsoxseg5ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8mf2_m))) +void vsoxseg5ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8mf4))) +void vsoxseg5ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8mf4_m))) +void vsoxseg5ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8mf8))) +void vsoxseg5ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8mf8_m))) +void vsoxseg5ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8m1))) +void vsoxseg6ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8m1_m))) +void vsoxseg6ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8mf2))) +void vsoxseg6ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8mf2_m))) +void vsoxseg6ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8mf4))) +void vsoxseg6ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8mf4_m))) +void vsoxseg6ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8mf8))) +void vsoxseg6ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8mf8_m))) +void vsoxseg6ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8m1))) +void vsoxseg6ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8m1_m))) +void vsoxseg6ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8mf2))) +void vsoxseg6ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8mf2_m))) +void vsoxseg6ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8mf4))) +void vsoxseg6ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8mf4_m))) +void vsoxseg6ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8mf8))) +void vsoxseg6ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8mf8_m))) +void vsoxseg6ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8m1))) +void vsoxseg7ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8m1_m))) +void vsoxseg7ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8mf2))) +void vsoxseg7ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8mf2_m))) +void vsoxseg7ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8mf4))) +void vsoxseg7ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8mf4_m))) +void vsoxseg7ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8mf8))) +void vsoxseg7ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8mf8_m))) +void vsoxseg7ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8m1))) +void vsoxseg7ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8m1_m))) +void vsoxseg7ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8mf2))) +void vsoxseg7ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8mf2_m))) +void vsoxseg7ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8mf4))) +void vsoxseg7ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8mf4_m))) +void vsoxseg7ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8mf8))) +void vsoxseg7ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8mf8_m))) +void vsoxseg7ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8m1))) +void vsoxseg8ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8m1_m))) +void vsoxseg8ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8mf2))) +void vsoxseg8ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8mf2_m))) +void vsoxseg8ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8mf4))) +void vsoxseg8ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8mf4_m))) +void vsoxseg8ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8mf8))) +void vsoxseg8ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8mf8_m))) +void vsoxseg8ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8m1))) +void vsoxseg8ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8m1_m))) +void vsoxseg8ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8mf2))) +void vsoxseg8ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8mf2_m))) +void vsoxseg8ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8mf4))) +void vsoxseg8ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8mf4_m))) +void vsoxseg8ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8mf8))) +void vsoxseg8ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8mf8_m))) +void vsoxseg8ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8m1))) +void vsoxseg2ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8m1_m))) +void vsoxseg2ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8mf2))) +void vsoxseg2ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8mf2_m))) +void vsoxseg2ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8mf4))) +void vsoxseg2ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8mf4_m))) +void vsoxseg2ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8mf8))) +void vsoxseg2ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8mf8_m))) +void vsoxseg2ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8m1))) +void vsoxseg2ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8m1_m))) +void vsoxseg2ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8mf2))) +void vsoxseg2ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8mf2_m))) +void vsoxseg2ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8mf4))) +void vsoxseg2ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8mf4_m))) +void vsoxseg2ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8mf8))) +void vsoxseg2ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8mf8_m))) +void vsoxseg2ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8m1))) +void vsoxseg3ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8m1_m))) +void vsoxseg3ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8mf2))) +void vsoxseg3ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8mf2_m))) +void vsoxseg3ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8mf4))) +void vsoxseg3ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8mf4_m))) +void vsoxseg3ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8mf8))) +void vsoxseg3ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8mf8_m))) +void vsoxseg3ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8m1))) +void vsoxseg3ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8m1_m))) +void vsoxseg3ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8mf2))) +void vsoxseg3ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8mf2_m))) +void vsoxseg3ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8mf4))) +void vsoxseg3ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8mf4_m))) +void vsoxseg3ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8mf8))) +void vsoxseg3ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8mf8_m))) +void vsoxseg3ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8m1))) +void vsoxseg4ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8m1_m))) +void vsoxseg4ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8mf2))) +void vsoxseg4ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8mf2_m))) +void vsoxseg4ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8mf4))) +void vsoxseg4ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8mf4_m))) +void vsoxseg4ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8mf8))) +void vsoxseg4ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8mf8_m))) +void vsoxseg4ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8m1))) +void vsoxseg4ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8m1_m))) +void vsoxseg4ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8mf2))) +void vsoxseg4ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8mf2_m))) +void vsoxseg4ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8mf4))) +void vsoxseg4ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8mf4_m))) +void vsoxseg4ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8mf8))) +void vsoxseg4ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8mf8_m))) +void vsoxseg4ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8m1))) +void vsoxseg5ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8m1_m))) +void vsoxseg5ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8mf2))) +void vsoxseg5ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8mf2_m))) +void vsoxseg5ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8mf4))) +void vsoxseg5ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8mf4_m))) +void vsoxseg5ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8mf8))) +void vsoxseg5ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8mf8_m))) +void vsoxseg5ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8m1))) +void vsoxseg5ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8m1_m))) +void vsoxseg5ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8mf2))) +void vsoxseg5ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8mf2_m))) +void vsoxseg5ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8mf4))) +void vsoxseg5ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8mf4_m))) +void vsoxseg5ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8mf8))) +void vsoxseg5ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8mf8_m))) +void vsoxseg5ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8m1))) +void vsoxseg6ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8m1_m))) +void vsoxseg6ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8mf2))) +void vsoxseg6ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8mf2_m))) +void vsoxseg6ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8mf4))) +void vsoxseg6ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8mf4_m))) +void vsoxseg6ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8mf8))) +void vsoxseg6ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8mf8_m))) +void vsoxseg6ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8m1))) +void vsoxseg6ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8m1_m))) +void vsoxseg6ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8mf2))) +void vsoxseg6ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8mf2_m))) +void vsoxseg6ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8mf4))) +void vsoxseg6ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8mf4_m))) +void vsoxseg6ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8mf8))) +void vsoxseg6ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8mf8_m))) +void vsoxseg6ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8m1))) +void vsoxseg7ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8m1_m))) +void vsoxseg7ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8mf2))) +void vsoxseg7ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8mf2_m))) +void vsoxseg7ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8mf4))) +void vsoxseg7ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8mf4_m))) +void vsoxseg7ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8mf8))) +void vsoxseg7ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8mf8_m))) +void vsoxseg7ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8m1))) +void vsoxseg7ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8m1_m))) +void vsoxseg7ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8mf2))) +void vsoxseg7ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8mf2_m))) +void vsoxseg7ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8mf4))) +void vsoxseg7ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8mf4_m))) +void vsoxseg7ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8mf8))) +void vsoxseg7ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8mf8_m))) +void vsoxseg7ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8m1))) +void vsoxseg8ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8m1_m))) +void vsoxseg8ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8mf2))) +void vsoxseg8ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8mf2_m))) +void vsoxseg8ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8mf4))) +void vsoxseg8ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8mf4_m))) +void vsoxseg8ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8mf8))) +void vsoxseg8ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8mf8_m))) +void vsoxseg8ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8m1))) +void vsoxseg8ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8m1_m))) +void vsoxseg8ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8mf2))) +void vsoxseg8ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8mf2_m))) +void vsoxseg8ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8mf4))) +void vsoxseg8ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8mf4_m))) +void vsoxseg8ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8mf8))) +void vsoxseg8ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8mf8_m))) +void vsoxseg8ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16m1))) +void vsoxseg2ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16m1_m))) +void vsoxseg2ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16m2))) +void vsoxseg2ei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16m2_m))) +void vsoxseg2ei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16m4))) +void vsoxseg2ei8(int16_t * op0, vuint8m2_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16m4_m))) +void vsoxseg2ei8(vbool4_t op0, int16_t * op1, vuint8m2_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16mf2))) +void vsoxseg2ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16mf2_m))) +void vsoxseg2ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16mf4))) +void vsoxseg2ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16mf4_m))) +void vsoxseg2ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16m1))) +void vsoxseg2ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16m1_m))) +void vsoxseg2ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16m2))) +void vsoxseg2ei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16m2_m))) +void vsoxseg2ei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16m4))) +void vsoxseg2ei8(uint16_t * op0, vuint8m2_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16m4_m))) +void vsoxseg2ei8(vbool4_t op0, uint16_t * op1, vuint8m2_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16mf2))) +void vsoxseg2ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16mf2_m))) +void vsoxseg2ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16mf4))) +void vsoxseg2ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16mf4_m))) +void vsoxseg2ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16m1))) +void vsoxseg3ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16m1_m))) +void vsoxseg3ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16m2))) +void vsoxseg3ei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16m2_m))) +void vsoxseg3ei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16mf2))) +void vsoxseg3ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16mf2_m))) +void vsoxseg3ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16mf4))) +void vsoxseg3ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16mf4_m))) +void vsoxseg3ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16m1))) +void vsoxseg3ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16m1_m))) +void vsoxseg3ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16m2))) +void vsoxseg3ei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16m2_m))) +void vsoxseg3ei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16mf2))) +void vsoxseg3ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16mf2_m))) +void vsoxseg3ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16mf4))) +void vsoxseg3ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16mf4_m))) +void vsoxseg3ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16m1))) +void vsoxseg4ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16m1_m))) +void vsoxseg4ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16m2))) +void vsoxseg4ei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16m2_m))) +void vsoxseg4ei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16mf2))) +void vsoxseg4ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16mf2_m))) +void vsoxseg4ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16mf4))) +void vsoxseg4ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16mf4_m))) +void vsoxseg4ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16m1))) +void vsoxseg4ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16m1_m))) +void vsoxseg4ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16m2))) +void vsoxseg4ei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16m2_m))) +void vsoxseg4ei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16mf2))) +void vsoxseg4ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16mf2_m))) +void vsoxseg4ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16mf4))) +void vsoxseg4ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16mf4_m))) +void vsoxseg4ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i16m1))) +void vsoxseg5ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i16m1_m))) +void vsoxseg5ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i16mf2))) +void vsoxseg5ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i16mf2_m))) +void vsoxseg5ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i16mf4))) +void vsoxseg5ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i16mf4_m))) +void vsoxseg5ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u16m1))) +void vsoxseg5ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u16m1_m))) +void vsoxseg5ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u16mf2))) +void vsoxseg5ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u16mf2_m))) +void vsoxseg5ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u16mf4))) +void vsoxseg5ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u16mf4_m))) +void vsoxseg5ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i16m1))) +void vsoxseg6ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i16m1_m))) +void vsoxseg6ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i16mf2))) +void vsoxseg6ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i16mf2_m))) +void vsoxseg6ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i16mf4))) +void vsoxseg6ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i16mf4_m))) +void vsoxseg6ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u16m1))) +void vsoxseg6ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u16m1_m))) +void vsoxseg6ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u16mf2))) +void vsoxseg6ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u16mf2_m))) +void vsoxseg6ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u16mf4))) +void vsoxseg6ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u16mf4_m))) +void vsoxseg6ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i16m1))) +void vsoxseg7ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i16m1_m))) +void vsoxseg7ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i16mf2))) +void vsoxseg7ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i16mf2_m))) +void vsoxseg7ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i16mf4))) +void vsoxseg7ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i16mf4_m))) +void vsoxseg7ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u16m1))) +void vsoxseg7ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u16m1_m))) +void vsoxseg7ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u16mf2))) +void vsoxseg7ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u16mf2_m))) +void vsoxseg7ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u16mf4))) +void vsoxseg7ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u16mf4_m))) +void vsoxseg7ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i16m1))) +void vsoxseg8ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i16m1_m))) +void vsoxseg8ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i16mf2))) +void vsoxseg8ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i16mf2_m))) +void vsoxseg8ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i16mf4))) +void vsoxseg8ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i16mf4_m))) +void vsoxseg8ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u16m1))) +void vsoxseg8ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u16m1_m))) +void vsoxseg8ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u16mf2))) +void vsoxseg8ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u16mf2_m))) +void vsoxseg8ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u16mf4))) +void vsoxseg8ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u16mf4_m))) +void vsoxseg8ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16m1))) +void vsoxseg2ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16m1_m))) +void vsoxseg2ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16m2))) +void vsoxseg2ei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16m2_m))) +void vsoxseg2ei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16m4))) +void vsoxseg2ei16(int16_t * op0, vuint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16m4_m))) +void vsoxseg2ei16(vbool4_t op0, int16_t * op1, vuint16m4_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16mf2))) +void vsoxseg2ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16mf2_m))) +void vsoxseg2ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16mf4))) +void vsoxseg2ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16mf4_m))) +void vsoxseg2ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16m1))) +void vsoxseg2ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16m1_m))) +void vsoxseg2ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16m2))) +void vsoxseg2ei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16m2_m))) +void vsoxseg2ei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16m4))) +void vsoxseg2ei16(uint16_t * op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16m4_m))) +void vsoxseg2ei16(vbool4_t op0, uint16_t * op1, vuint16m4_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16mf2))) +void vsoxseg2ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16mf2_m))) +void vsoxseg2ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16mf4))) +void vsoxseg2ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16mf4_m))) +void vsoxseg2ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16m1))) +void vsoxseg3ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16m1_m))) +void vsoxseg3ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16m2))) +void vsoxseg3ei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16m2_m))) +void vsoxseg3ei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16mf2))) +void vsoxseg3ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16mf2_m))) +void vsoxseg3ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16mf4))) +void vsoxseg3ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16mf4_m))) +void vsoxseg3ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16m1))) +void vsoxseg3ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16m1_m))) +void vsoxseg3ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16m2))) +void vsoxseg3ei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16m2_m))) +void vsoxseg3ei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16mf2))) +void vsoxseg3ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16mf2_m))) +void vsoxseg3ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16mf4))) +void vsoxseg3ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16mf4_m))) +void vsoxseg3ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16m1))) +void vsoxseg4ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16m1_m))) +void vsoxseg4ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16m2))) +void vsoxseg4ei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16m2_m))) +void vsoxseg4ei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16mf2))) +void vsoxseg4ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16mf2_m))) +void vsoxseg4ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16mf4))) +void vsoxseg4ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16mf4_m))) +void vsoxseg4ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16m1))) +void vsoxseg4ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16m1_m))) +void vsoxseg4ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16m2))) +void vsoxseg4ei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16m2_m))) +void vsoxseg4ei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16mf2))) +void vsoxseg4ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16mf2_m))) +void vsoxseg4ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16mf4))) +void vsoxseg4ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16mf4_m))) +void vsoxseg4ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i16m1))) +void vsoxseg5ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i16m1_m))) +void vsoxseg5ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i16mf2))) +void vsoxseg5ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i16mf2_m))) +void vsoxseg5ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i16mf4))) +void vsoxseg5ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i16mf4_m))) +void vsoxseg5ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u16m1))) +void vsoxseg5ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u16m1_m))) +void vsoxseg5ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u16mf2))) +void vsoxseg5ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u16mf2_m))) +void vsoxseg5ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u16mf4))) +void vsoxseg5ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u16mf4_m))) +void vsoxseg5ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i16m1))) +void vsoxseg6ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i16m1_m))) +void vsoxseg6ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i16mf2))) +void vsoxseg6ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i16mf2_m))) +void vsoxseg6ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i16mf4))) +void vsoxseg6ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i16mf4_m))) +void vsoxseg6ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u16m1))) +void vsoxseg6ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u16m1_m))) +void vsoxseg6ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u16mf2))) +void vsoxseg6ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u16mf2_m))) +void vsoxseg6ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u16mf4))) +void vsoxseg6ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u16mf4_m))) +void vsoxseg6ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i16m1))) +void vsoxseg7ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i16m1_m))) +void vsoxseg7ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i16mf2))) +void vsoxseg7ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i16mf2_m))) +void vsoxseg7ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i16mf4))) +void vsoxseg7ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i16mf4_m))) +void vsoxseg7ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u16m1))) +void vsoxseg7ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u16m1_m))) +void vsoxseg7ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u16mf2))) +void vsoxseg7ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u16mf2_m))) +void vsoxseg7ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u16mf4))) +void vsoxseg7ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u16mf4_m))) +void vsoxseg7ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i16m1))) +void vsoxseg8ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i16m1_m))) +void vsoxseg8ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i16mf2))) +void vsoxseg8ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i16mf2_m))) +void vsoxseg8ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i16mf4))) +void vsoxseg8ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i16mf4_m))) +void vsoxseg8ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u16m1))) +void vsoxseg8ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u16m1_m))) +void vsoxseg8ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u16mf2))) +void vsoxseg8ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u16mf2_m))) +void vsoxseg8ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u16mf4))) +void vsoxseg8ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u16mf4_m))) +void vsoxseg8ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16m1))) +void vsoxseg2ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16m1_m))) +void vsoxseg2ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16m2))) +void vsoxseg2ei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16m2_m))) +void vsoxseg2ei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16m4))) +void vsoxseg2ei32(int16_t * op0, vuint32m8_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16m4_m))) +void vsoxseg2ei32(vbool4_t op0, int16_t * op1, vuint32m8_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16mf2))) +void vsoxseg2ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16mf2_m))) +void vsoxseg2ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16mf4))) +void vsoxseg2ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16mf4_m))) +void vsoxseg2ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16m1))) +void vsoxseg2ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16m1_m))) +void vsoxseg2ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16m2))) +void vsoxseg2ei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16m2_m))) +void vsoxseg2ei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16m4))) +void vsoxseg2ei32(uint16_t * op0, vuint32m8_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16m4_m))) +void vsoxseg2ei32(vbool4_t op0, uint16_t * op1, vuint32m8_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16mf2))) +void vsoxseg2ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16mf2_m))) +void vsoxseg2ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16mf4))) +void vsoxseg2ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16mf4_m))) +void vsoxseg2ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16m1))) +void vsoxseg3ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16m1_m))) +void vsoxseg3ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16m2))) +void vsoxseg3ei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16m2_m))) +void vsoxseg3ei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16mf2))) +void vsoxseg3ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16mf2_m))) +void vsoxseg3ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16mf4))) +void vsoxseg3ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16mf4_m))) +void vsoxseg3ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16m1))) +void vsoxseg3ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16m1_m))) +void vsoxseg3ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16m2))) +void vsoxseg3ei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16m2_m))) +void vsoxseg3ei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16mf2))) +void vsoxseg3ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16mf2_m))) +void vsoxseg3ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16mf4))) +void vsoxseg3ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16mf4_m))) +void vsoxseg3ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16m1))) +void vsoxseg4ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16m1_m))) +void vsoxseg4ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16m2))) +void vsoxseg4ei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16m2_m))) +void vsoxseg4ei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16mf2))) +void vsoxseg4ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16mf2_m))) +void vsoxseg4ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16mf4))) +void vsoxseg4ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16mf4_m))) +void vsoxseg4ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16m1))) +void vsoxseg4ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16m1_m))) +void vsoxseg4ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16m2))) +void vsoxseg4ei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16m2_m))) +void vsoxseg4ei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16mf2))) +void vsoxseg4ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16mf2_m))) +void vsoxseg4ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16mf4))) +void vsoxseg4ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16mf4_m))) +void vsoxseg4ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i16m1))) +void vsoxseg5ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i16m1_m))) +void vsoxseg5ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i16mf2))) +void vsoxseg5ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i16mf2_m))) +void vsoxseg5ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i16mf4))) +void vsoxseg5ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i16mf4_m))) +void vsoxseg5ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u16m1))) +void vsoxseg5ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u16m1_m))) +void vsoxseg5ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u16mf2))) +void vsoxseg5ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u16mf2_m))) +void vsoxseg5ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u16mf4))) +void vsoxseg5ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u16mf4_m))) +void vsoxseg5ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i16m1))) +void vsoxseg6ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i16m1_m))) +void vsoxseg6ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i16mf2))) +void vsoxseg6ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i16mf2_m))) +void vsoxseg6ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i16mf4))) +void vsoxseg6ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i16mf4_m))) +void vsoxseg6ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u16m1))) +void vsoxseg6ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u16m1_m))) +void vsoxseg6ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u16mf2))) +void vsoxseg6ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u16mf2_m))) +void vsoxseg6ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u16mf4))) +void vsoxseg6ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u16mf4_m))) +void vsoxseg6ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i16m1))) +void vsoxseg7ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i16m1_m))) +void vsoxseg7ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i16mf2))) +void vsoxseg7ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i16mf2_m))) +void vsoxseg7ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i16mf4))) +void vsoxseg7ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i16mf4_m))) +void vsoxseg7ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u16m1))) +void vsoxseg7ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u16m1_m))) +void vsoxseg7ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u16mf2))) +void vsoxseg7ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u16mf2_m))) +void vsoxseg7ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u16mf4))) +void vsoxseg7ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u16mf4_m))) +void vsoxseg7ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i16m1))) +void vsoxseg8ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i16m1_m))) +void vsoxseg8ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i16mf2))) +void vsoxseg8ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i16mf2_m))) +void vsoxseg8ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i16mf4))) +void vsoxseg8ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i16mf4_m))) +void vsoxseg8ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u16m1))) +void vsoxseg8ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u16m1_m))) +void vsoxseg8ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u16mf2))) +void vsoxseg8ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u16mf2_m))) +void vsoxseg8ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u16mf4))) +void vsoxseg8ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u16mf4_m))) +void vsoxseg8ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16m1))) +void vsoxseg2ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16m1_m))) +void vsoxseg2ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16m2))) +void vsoxseg2ei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16m2_m))) +void vsoxseg2ei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16mf2))) +void vsoxseg2ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16mf2_m))) +void vsoxseg2ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16mf4))) +void vsoxseg2ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16mf4_m))) +void vsoxseg2ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16m1))) +void vsoxseg2ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16m1_m))) +void vsoxseg2ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16m2))) +void vsoxseg2ei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16m2_m))) +void vsoxseg2ei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16mf2))) +void vsoxseg2ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16mf2_m))) +void vsoxseg2ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16mf4))) +void vsoxseg2ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16mf4_m))) +void vsoxseg2ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16m1))) +void vsoxseg3ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16m1_m))) +void vsoxseg3ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16m2))) +void vsoxseg3ei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16m2_m))) +void vsoxseg3ei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16mf2))) +void vsoxseg3ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16mf2_m))) +void vsoxseg3ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16mf4))) +void vsoxseg3ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16mf4_m))) +void vsoxseg3ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16m1))) +void vsoxseg3ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16m1_m))) +void vsoxseg3ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16m2))) +void vsoxseg3ei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16m2_m))) +void vsoxseg3ei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16mf2))) +void vsoxseg3ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16mf2_m))) +void vsoxseg3ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16mf4))) +void vsoxseg3ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16mf4_m))) +void vsoxseg3ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16m1))) +void vsoxseg4ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16m1_m))) +void vsoxseg4ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16m2))) +void vsoxseg4ei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16m2_m))) +void vsoxseg4ei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16mf2))) +void vsoxseg4ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16mf2_m))) +void vsoxseg4ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16mf4))) +void vsoxseg4ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16mf4_m))) +void vsoxseg4ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16m1))) +void vsoxseg4ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16m1_m))) +void vsoxseg4ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16m2))) +void vsoxseg4ei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16m2_m))) +void vsoxseg4ei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16mf2))) +void vsoxseg4ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16mf2_m))) +void vsoxseg4ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16mf4))) +void vsoxseg4ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16mf4_m))) +void vsoxseg4ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i16m1))) +void vsoxseg5ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i16m1_m))) +void vsoxseg5ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i16mf2))) +void vsoxseg5ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i16mf2_m))) +void vsoxseg5ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i16mf4))) +void vsoxseg5ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i16mf4_m))) +void vsoxseg5ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u16m1))) +void vsoxseg5ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u16m1_m))) +void vsoxseg5ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u16mf2))) +void vsoxseg5ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u16mf2_m))) +void vsoxseg5ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u16mf4))) +void vsoxseg5ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u16mf4_m))) +void vsoxseg5ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i16m1))) +void vsoxseg6ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i16m1_m))) +void vsoxseg6ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i16mf2))) +void vsoxseg6ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i16mf2_m))) +void vsoxseg6ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i16mf4))) +void vsoxseg6ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i16mf4_m))) +void vsoxseg6ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u16m1))) +void vsoxseg6ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u16m1_m))) +void vsoxseg6ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u16mf2))) +void vsoxseg6ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u16mf2_m))) +void vsoxseg6ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u16mf4))) +void vsoxseg6ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u16mf4_m))) +void vsoxseg6ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i16m1))) +void vsoxseg7ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i16m1_m))) +void vsoxseg7ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i16mf2))) +void vsoxseg7ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i16mf2_m))) +void vsoxseg7ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i16mf4))) +void vsoxseg7ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i16mf4_m))) +void vsoxseg7ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u16m1))) +void vsoxseg7ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u16m1_m))) +void vsoxseg7ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u16mf2))) +void vsoxseg7ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u16mf2_m))) +void vsoxseg7ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u16mf4))) +void vsoxseg7ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u16mf4_m))) +void vsoxseg7ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i16m1))) +void vsoxseg8ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i16m1_m))) +void vsoxseg8ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i16mf2))) +void vsoxseg8ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i16mf2_m))) +void vsoxseg8ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i16mf4))) +void vsoxseg8ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i16mf4_m))) +void vsoxseg8ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u16m1))) +void vsoxseg8ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u16m1_m))) +void vsoxseg8ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u16mf2))) +void vsoxseg8ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u16mf2_m))) +void vsoxseg8ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u16mf4))) +void vsoxseg8ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u16mf4_m))) +void vsoxseg8ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32m1))) +void vsoxseg2ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32m1_m))) +void vsoxseg2ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32m2))) +void vsoxseg2ei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32m2_m))) +void vsoxseg2ei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32m4))) +void vsoxseg2ei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32m4_m))) +void vsoxseg2ei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32mf2))) +void vsoxseg2ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32mf2_m))) +void vsoxseg2ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32m1))) +void vsoxseg2ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32m1_m))) +void vsoxseg2ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32m2))) +void vsoxseg2ei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32m2_m))) +void vsoxseg2ei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32m4))) +void vsoxseg2ei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32m4_m))) +void vsoxseg2ei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32mf2))) +void vsoxseg2ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32mf2_m))) +void vsoxseg2ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i32m1))) +void vsoxseg3ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i32m1_m))) +void vsoxseg3ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i32m2))) +void vsoxseg3ei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i32m2_m))) +void vsoxseg3ei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i32mf2))) +void vsoxseg3ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i32mf2_m))) +void vsoxseg3ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u32m1))) +void vsoxseg3ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u32m1_m))) +void vsoxseg3ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u32m2))) +void vsoxseg3ei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u32m2_m))) +void vsoxseg3ei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u32mf2))) +void vsoxseg3ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u32mf2_m))) +void vsoxseg3ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i32m1))) +void vsoxseg4ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i32m1_m))) +void vsoxseg4ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i32m2))) +void vsoxseg4ei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i32m2_m))) +void vsoxseg4ei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i32mf2))) +void vsoxseg4ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i32mf2_m))) +void vsoxseg4ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u32m1))) +void vsoxseg4ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u32m1_m))) +void vsoxseg4ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u32m2))) +void vsoxseg4ei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u32m2_m))) +void vsoxseg4ei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u32mf2))) +void vsoxseg4ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u32mf2_m))) +void vsoxseg4ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i32m1))) +void vsoxseg5ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i32m1_m))) +void vsoxseg5ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i32mf2))) +void vsoxseg5ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i32mf2_m))) +void vsoxseg5ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u32m1))) +void vsoxseg5ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u32m1_m))) +void vsoxseg5ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u32mf2))) +void vsoxseg5ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u32mf2_m))) +void vsoxseg5ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i32m1))) +void vsoxseg6ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i32m1_m))) +void vsoxseg6ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i32mf2))) +void vsoxseg6ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i32mf2_m))) +void vsoxseg6ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u32m1))) +void vsoxseg6ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u32m1_m))) +void vsoxseg6ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u32mf2))) +void vsoxseg6ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u32mf2_m))) +void vsoxseg6ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i32m1))) +void vsoxseg7ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i32m1_m))) +void vsoxseg7ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i32mf2))) +void vsoxseg7ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i32mf2_m))) +void vsoxseg7ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u32m1))) +void vsoxseg7ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u32m1_m))) +void vsoxseg7ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u32mf2))) +void vsoxseg7ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u32mf2_m))) +void vsoxseg7ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i32m1))) +void vsoxseg8ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i32m1_m))) +void vsoxseg8ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i32mf2))) +void vsoxseg8ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i32mf2_m))) +void vsoxseg8ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u32m1))) +void vsoxseg8ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u32m1_m))) +void vsoxseg8ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u32mf2))) +void vsoxseg8ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u32mf2_m))) +void vsoxseg8ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32m1))) +void vsoxseg2ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32m1_m))) +void vsoxseg2ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32m2))) +void vsoxseg2ei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32m2_m))) +void vsoxseg2ei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32m4))) +void vsoxseg2ei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32m4_m))) +void vsoxseg2ei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32mf2))) +void vsoxseg2ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32mf2_m))) +void vsoxseg2ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32m1))) +void vsoxseg2ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32m1_m))) +void vsoxseg2ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32m2))) +void vsoxseg2ei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32m2_m))) +void vsoxseg2ei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32m4))) +void vsoxseg2ei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32m4_m))) +void vsoxseg2ei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32mf2))) +void vsoxseg2ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32mf2_m))) +void vsoxseg2ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i32m1))) +void vsoxseg3ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i32m1_m))) +void vsoxseg3ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i32m2))) +void vsoxseg3ei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i32m2_m))) +void vsoxseg3ei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i32mf2))) +void vsoxseg3ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i32mf2_m))) +void vsoxseg3ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u32m1))) +void vsoxseg3ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u32m1_m))) +void vsoxseg3ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u32m2))) +void vsoxseg3ei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u32m2_m))) +void vsoxseg3ei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u32mf2))) +void vsoxseg3ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u32mf2_m))) +void vsoxseg3ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i32m1))) +void vsoxseg4ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i32m1_m))) +void vsoxseg4ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i32m2))) +void vsoxseg4ei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i32m2_m))) +void vsoxseg4ei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i32mf2))) +void vsoxseg4ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i32mf2_m))) +void vsoxseg4ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u32m1))) +void vsoxseg4ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u32m1_m))) +void vsoxseg4ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u32m2))) +void vsoxseg4ei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u32m2_m))) +void vsoxseg4ei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u32mf2))) +void vsoxseg4ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u32mf2_m))) +void vsoxseg4ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i32m1))) +void vsoxseg5ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i32m1_m))) +void vsoxseg5ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i32mf2))) +void vsoxseg5ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i32mf2_m))) +void vsoxseg5ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u32m1))) +void vsoxseg5ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u32m1_m))) +void vsoxseg5ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u32mf2))) +void vsoxseg5ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u32mf2_m))) +void vsoxseg5ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i32m1))) +void vsoxseg6ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i32m1_m))) +void vsoxseg6ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i32mf2))) +void vsoxseg6ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i32mf2_m))) +void vsoxseg6ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u32m1))) +void vsoxseg6ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u32m1_m))) +void vsoxseg6ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u32mf2))) +void vsoxseg6ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u32mf2_m))) +void vsoxseg6ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i32m1))) +void vsoxseg7ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i32m1_m))) +void vsoxseg7ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i32mf2))) +void vsoxseg7ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i32mf2_m))) +void vsoxseg7ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u32m1))) +void vsoxseg7ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u32m1_m))) +void vsoxseg7ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u32mf2))) +void vsoxseg7ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u32mf2_m))) +void vsoxseg7ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i32m1))) +void vsoxseg8ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i32m1_m))) +void vsoxseg8ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i32mf2))) +void vsoxseg8ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i32mf2_m))) +void vsoxseg8ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u32m1))) +void vsoxseg8ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u32m1_m))) +void vsoxseg8ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u32mf2))) +void vsoxseg8ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u32mf2_m))) +void vsoxseg8ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32m1))) +void vsoxseg2ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32m1_m))) +void vsoxseg2ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32m2))) +void vsoxseg2ei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32m2_m))) +void vsoxseg2ei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32m4))) +void vsoxseg2ei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32m4_m))) +void vsoxseg2ei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32mf2))) +void vsoxseg2ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32mf2_m))) +void vsoxseg2ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32m1))) +void vsoxseg2ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32m1_m))) +void vsoxseg2ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32m2))) +void vsoxseg2ei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32m2_m))) +void vsoxseg2ei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32m4))) +void vsoxseg2ei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32m4_m))) +void vsoxseg2ei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32mf2))) +void vsoxseg2ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32mf2_m))) +void vsoxseg2ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i32m1))) +void vsoxseg3ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i32m1_m))) +void vsoxseg3ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i32m2))) +void vsoxseg3ei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i32m2_m))) +void vsoxseg3ei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i32mf2))) +void vsoxseg3ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i32mf2_m))) +void vsoxseg3ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u32m1))) +void vsoxseg3ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u32m1_m))) +void vsoxseg3ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u32m2))) +void vsoxseg3ei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u32m2_m))) +void vsoxseg3ei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u32mf2))) +void vsoxseg3ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u32mf2_m))) +void vsoxseg3ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i32m1))) +void vsoxseg4ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i32m1_m))) +void vsoxseg4ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i32m2))) +void vsoxseg4ei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i32m2_m))) +void vsoxseg4ei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i32mf2))) +void vsoxseg4ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i32mf2_m))) +void vsoxseg4ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u32m1))) +void vsoxseg4ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u32m1_m))) +void vsoxseg4ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u32m2))) +void vsoxseg4ei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u32m2_m))) +void vsoxseg4ei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u32mf2))) +void vsoxseg4ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u32mf2_m))) +void vsoxseg4ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i32m1))) +void vsoxseg5ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i32m1_m))) +void vsoxseg5ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i32mf2))) +void vsoxseg5ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i32mf2_m))) +void vsoxseg5ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u32m1))) +void vsoxseg5ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u32m1_m))) +void vsoxseg5ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u32mf2))) +void vsoxseg5ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u32mf2_m))) +void vsoxseg5ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i32m1))) +void vsoxseg6ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i32m1_m))) +void vsoxseg6ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i32mf2))) +void vsoxseg6ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i32mf2_m))) +void vsoxseg6ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u32m1))) +void vsoxseg6ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u32m1_m))) +void vsoxseg6ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u32mf2))) +void vsoxseg6ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u32mf2_m))) +void vsoxseg6ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i32m1))) +void vsoxseg7ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i32m1_m))) +void vsoxseg7ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i32mf2))) +void vsoxseg7ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i32mf2_m))) +void vsoxseg7ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u32m1))) +void vsoxseg7ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u32m1_m))) +void vsoxseg7ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u32mf2))) +void vsoxseg7ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u32mf2_m))) +void vsoxseg7ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i32m1))) +void vsoxseg8ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i32m1_m))) +void vsoxseg8ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i32mf2))) +void vsoxseg8ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i32mf2_m))) +void vsoxseg8ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u32m1))) +void vsoxseg8ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u32m1_m))) +void vsoxseg8ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u32mf2))) +void vsoxseg8ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u32mf2_m))) +void vsoxseg8ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32m1))) +void vsoxseg2ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32m1_m))) +void vsoxseg2ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32m2))) +void vsoxseg2ei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32m2_m))) +void vsoxseg2ei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32m4))) +void vsoxseg2ei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32m4_m))) +void vsoxseg2ei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32mf2))) +void vsoxseg2ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32mf2_m))) +void vsoxseg2ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32m1))) +void vsoxseg2ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32m1_m))) +void vsoxseg2ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32m2))) +void vsoxseg2ei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32m2_m))) +void vsoxseg2ei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32m4))) +void vsoxseg2ei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32m4_m))) +void vsoxseg2ei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32mf2))) +void vsoxseg2ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32mf2_m))) +void vsoxseg2ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i32m1))) +void vsoxseg3ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i32m1_m))) +void vsoxseg3ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i32m2))) +void vsoxseg3ei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i32m2_m))) +void vsoxseg3ei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i32mf2))) +void vsoxseg3ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i32mf2_m))) +void vsoxseg3ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u32m1))) +void vsoxseg3ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u32m1_m))) +void vsoxseg3ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u32m2))) +void vsoxseg3ei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u32m2_m))) +void vsoxseg3ei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u32mf2))) +void vsoxseg3ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u32mf2_m))) +void vsoxseg3ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i32m1))) +void vsoxseg4ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i32m1_m))) +void vsoxseg4ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i32m2))) +void vsoxseg4ei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i32m2_m))) +void vsoxseg4ei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i32mf2))) +void vsoxseg4ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i32mf2_m))) +void vsoxseg4ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u32m1))) +void vsoxseg4ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u32m1_m))) +void vsoxseg4ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u32m2))) +void vsoxseg4ei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u32m2_m))) +void vsoxseg4ei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u32mf2))) +void vsoxseg4ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u32mf2_m))) +void vsoxseg4ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i32m1))) +void vsoxseg5ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i32m1_m))) +void vsoxseg5ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i32mf2))) +void vsoxseg5ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i32mf2_m))) +void vsoxseg5ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u32m1))) +void vsoxseg5ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u32m1_m))) +void vsoxseg5ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u32mf2))) +void vsoxseg5ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u32mf2_m))) +void vsoxseg5ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i32m1))) +void vsoxseg6ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i32m1_m))) +void vsoxseg6ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i32mf2))) +void vsoxseg6ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i32mf2_m))) +void vsoxseg6ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u32m1))) +void vsoxseg6ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u32m1_m))) +void vsoxseg6ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u32mf2))) +void vsoxseg6ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u32mf2_m))) +void vsoxseg6ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i32m1))) +void vsoxseg7ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i32m1_m))) +void vsoxseg7ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i32mf2))) +void vsoxseg7ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i32mf2_m))) +void vsoxseg7ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u32m1))) +void vsoxseg7ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u32m1_m))) +void vsoxseg7ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u32mf2))) +void vsoxseg7ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u32mf2_m))) +void vsoxseg7ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i32m1))) +void vsoxseg8ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i32m1_m))) +void vsoxseg8ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i32mf2))) +void vsoxseg8ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i32mf2_m))) +void vsoxseg8ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u32m1))) +void vsoxseg8ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u32m1_m))) +void vsoxseg8ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u32mf2))) +void vsoxseg8ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u32mf2_m))) +void vsoxseg8ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i64m1))) +void vsoxseg2ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i64m1_m))) +void vsoxseg2ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i64m2))) +void vsoxseg2ei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i64m2_m))) +void vsoxseg2ei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i64m4))) +void vsoxseg2ei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i64m4_m))) +void vsoxseg2ei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u64m1))) +void vsoxseg2ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u64m1_m))) +void vsoxseg2ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u64m2))) +void vsoxseg2ei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u64m2_m))) +void vsoxseg2ei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u64m4))) +void vsoxseg2ei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u64m4_m))) +void vsoxseg2ei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i64m1))) +void vsoxseg3ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i64m1_m))) +void vsoxseg3ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i64m2))) +void vsoxseg3ei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i64m2_m))) +void vsoxseg3ei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u64m1))) +void vsoxseg3ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u64m1_m))) +void vsoxseg3ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u64m2))) +void vsoxseg3ei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u64m2_m))) +void vsoxseg3ei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i64m1))) +void vsoxseg4ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i64m1_m))) +void vsoxseg4ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i64m2))) +void vsoxseg4ei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i64m2_m))) +void vsoxseg4ei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u64m1))) +void vsoxseg4ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u64m1_m))) +void vsoxseg4ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u64m2))) +void vsoxseg4ei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u64m2_m))) +void vsoxseg4ei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i64m1))) +void vsoxseg5ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i64m1_m))) +void vsoxseg5ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u64m1))) +void vsoxseg5ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u64m1_m))) +void vsoxseg5ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i64m1))) +void vsoxseg6ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i64m1_m))) +void vsoxseg6ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u64m1))) +void vsoxseg6ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u64m1_m))) +void vsoxseg6ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i64m1))) +void vsoxseg7ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i64m1_m))) +void vsoxseg7ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u64m1))) +void vsoxseg7ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u64m1_m))) +void vsoxseg7ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i64m1))) +void vsoxseg8ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i64m1_m))) +void vsoxseg8ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u64m1))) +void vsoxseg8ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u64m1_m))) +void vsoxseg8ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i64m1))) +void vsoxseg2ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i64m1_m))) +void vsoxseg2ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i64m2))) +void vsoxseg2ei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i64m2_m))) +void vsoxseg2ei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i64m4))) +void vsoxseg2ei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i64m4_m))) +void vsoxseg2ei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u64m1))) +void vsoxseg2ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u64m1_m))) +void vsoxseg2ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u64m2))) +void vsoxseg2ei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u64m2_m))) +void vsoxseg2ei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u64m4))) +void vsoxseg2ei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u64m4_m))) +void vsoxseg2ei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i64m1))) +void vsoxseg3ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i64m1_m))) +void vsoxseg3ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i64m2))) +void vsoxseg3ei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i64m2_m))) +void vsoxseg3ei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u64m1))) +void vsoxseg3ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u64m1_m))) +void vsoxseg3ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u64m2))) +void vsoxseg3ei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u64m2_m))) +void vsoxseg3ei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i64m1))) +void vsoxseg4ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i64m1_m))) +void vsoxseg4ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i64m2))) +void vsoxseg4ei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i64m2_m))) +void vsoxseg4ei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u64m1))) +void vsoxseg4ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u64m1_m))) +void vsoxseg4ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u64m2))) +void vsoxseg4ei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u64m2_m))) +void vsoxseg4ei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i64m1))) +void vsoxseg5ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i64m1_m))) +void vsoxseg5ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u64m1))) +void vsoxseg5ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u64m1_m))) +void vsoxseg5ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i64m1))) +void vsoxseg6ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i64m1_m))) +void vsoxseg6ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u64m1))) +void vsoxseg6ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u64m1_m))) +void vsoxseg6ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i64m1))) +void vsoxseg7ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i64m1_m))) +void vsoxseg7ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u64m1))) +void vsoxseg7ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u64m1_m))) +void vsoxseg7ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i64m1))) +void vsoxseg8ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i64m1_m))) +void vsoxseg8ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u64m1))) +void vsoxseg8ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u64m1_m))) +void vsoxseg8ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i64m1))) +void vsoxseg2ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i64m1_m))) +void vsoxseg2ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i64m2))) +void vsoxseg2ei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i64m2_m))) +void vsoxseg2ei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i64m4))) +void vsoxseg2ei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i64m4_m))) +void vsoxseg2ei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u64m1))) +void vsoxseg2ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u64m1_m))) +void vsoxseg2ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u64m2))) +void vsoxseg2ei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u64m2_m))) +void vsoxseg2ei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u64m4))) +void vsoxseg2ei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u64m4_m))) +void vsoxseg2ei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i64m1))) +void vsoxseg3ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i64m1_m))) +void vsoxseg3ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i64m2))) +void vsoxseg3ei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i64m2_m))) +void vsoxseg3ei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u64m1))) +void vsoxseg3ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u64m1_m))) +void vsoxseg3ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u64m2))) +void vsoxseg3ei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u64m2_m))) +void vsoxseg3ei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i64m1))) +void vsoxseg4ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i64m1_m))) +void vsoxseg4ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i64m2))) +void vsoxseg4ei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i64m2_m))) +void vsoxseg4ei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u64m1))) +void vsoxseg4ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u64m1_m))) +void vsoxseg4ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u64m2))) +void vsoxseg4ei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u64m2_m))) +void vsoxseg4ei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i64m1))) +void vsoxseg5ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i64m1_m))) +void vsoxseg5ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u64m1))) +void vsoxseg5ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u64m1_m))) +void vsoxseg5ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i64m1))) +void vsoxseg6ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i64m1_m))) +void vsoxseg6ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u64m1))) +void vsoxseg6ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u64m1_m))) +void vsoxseg6ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i64m1))) +void vsoxseg7ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i64m1_m))) +void vsoxseg7ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u64m1))) +void vsoxseg7ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u64m1_m))) +void vsoxseg7ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i64m1))) +void vsoxseg8ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i64m1_m))) +void vsoxseg8ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u64m1))) +void vsoxseg8ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u64m1_m))) +void vsoxseg8ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i64m1))) +void vsoxseg2ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i64m1_m))) +void vsoxseg2ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i64m2))) +void vsoxseg2ei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i64m2_m))) +void vsoxseg2ei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i64m4))) +void vsoxseg2ei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i64m4_m))) +void vsoxseg2ei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u64m1))) +void vsoxseg2ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u64m1_m))) +void vsoxseg2ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u64m2))) +void vsoxseg2ei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u64m2_m))) +void vsoxseg2ei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u64m4))) +void vsoxseg2ei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u64m4_m))) +void vsoxseg2ei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i64m1))) +void vsoxseg3ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i64m1_m))) +void vsoxseg3ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i64m2))) +void vsoxseg3ei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i64m2_m))) +void vsoxseg3ei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u64m1))) +void vsoxseg3ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u64m1_m))) +void vsoxseg3ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u64m2))) +void vsoxseg3ei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u64m2_m))) +void vsoxseg3ei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i64m1))) +void vsoxseg4ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i64m1_m))) +void vsoxseg4ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i64m2))) +void vsoxseg4ei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i64m2_m))) +void vsoxseg4ei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u64m1))) +void vsoxseg4ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u64m1_m))) +void vsoxseg4ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u64m2))) +void vsoxseg4ei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u64m2_m))) +void vsoxseg4ei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i64m1))) +void vsoxseg5ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i64m1_m))) +void vsoxseg5ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u64m1))) +void vsoxseg5ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u64m1_m))) +void vsoxseg5ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i64m1))) +void vsoxseg6ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i64m1_m))) +void vsoxseg6ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u64m1))) +void vsoxseg6ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u64m1_m))) +void vsoxseg6ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i64m1))) +void vsoxseg7ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i64m1_m))) +void vsoxseg7ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u64m1))) +void vsoxseg7ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u64m1_m))) +void vsoxseg7ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i64m1))) +void vsoxseg8ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i64m1_m))) +void vsoxseg8ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u64m1))) +void vsoxseg8ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u64m1_m))) +void vsoxseg8ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_i8m1_m))) +void vlseg2e8(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_i8m2_m))) +void vlseg2e8(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_i8m4_m))) +void vlseg2e8(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_i8mf2_m))) +void vlseg2e8(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_i8mf4_m))) +void vlseg2e8(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_i8mf8_m))) +void vlseg2e8(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_u8m1_m))) +void vlseg2e8(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_u8m2_m))) +void vlseg2e8(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_u8m4_m))) +void vlseg2e8(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_u8mf2_m))) +void vlseg2e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_u8mf4_m))) +void vlseg2e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_u8mf8_m))) +void vlseg2e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_i8m1_m))) +void vlseg3e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_i8m2_m))) +void vlseg3e8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_i8mf2_m))) +void vlseg3e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_i8mf4_m))) +void vlseg3e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_i8mf8_m))) +void vlseg3e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_u8m1_m))) +void vlseg3e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_u8m2_m))) +void vlseg3e8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_u8mf2_m))) +void vlseg3e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_u8mf4_m))) +void vlseg3e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_u8mf8_m))) +void vlseg3e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_i8m1_m))) +void vlseg4e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_i8m2_m))) +void vlseg4e8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_i8mf2_m))) +void vlseg4e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_i8mf4_m))) +void vlseg4e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_i8mf8_m))) +void vlseg4e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_u8m1_m))) +void vlseg4e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_u8m2_m))) +void vlseg4e8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_u8mf2_m))) +void vlseg4e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_u8mf4_m))) +void vlseg4e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_u8mf8_m))) +void vlseg4e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_i8m1_m))) +void vlseg5e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_i8mf2_m))) +void vlseg5e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_i8mf4_m))) +void vlseg5e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_i8mf8_m))) +void vlseg5e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_u8m1_m))) +void vlseg5e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_u8mf2_m))) +void vlseg5e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_u8mf4_m))) +void vlseg5e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_u8mf8_m))) +void vlseg5e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_i8m1_m))) +void vlseg6e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_i8mf2_m))) +void vlseg6e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_i8mf4_m))) +void vlseg6e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_i8mf8_m))) +void vlseg6e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_u8m1_m))) +void vlseg6e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_u8mf2_m))) +void vlseg6e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_u8mf4_m))) +void vlseg6e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_u8mf8_m))) +void vlseg6e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_i8m1_m))) +void vlseg7e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_i8mf2_m))) +void vlseg7e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_i8mf4_m))) +void vlseg7e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_i8mf8_m))) +void vlseg7e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_u8m1_m))) +void vlseg7e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_u8mf2_m))) +void vlseg7e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_u8mf4_m))) +void vlseg7e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_u8mf8_m))) +void vlseg7e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_i8m1_m))) +void vlseg8e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_i8mf2_m))) +void vlseg8e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_i8mf4_m))) +void vlseg8e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_i8mf8_m))) +void vlseg8e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_u8m1_m))) +void vlseg8e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_u8mf2_m))) +void vlseg8e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_u8mf4_m))) +void vlseg8e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_u8mf8_m))) +void vlseg8e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_i16m1_m))) +void vlseg2e16(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_i16m2_m))) +void vlseg2e16(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_i16m4_m))) +void vlseg2e16(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_i16mf2_m))) +void vlseg2e16(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_i16mf4_m))) +void vlseg2e16(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_u16m1_m))) +void vlseg2e16(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_u16m2_m))) +void vlseg2e16(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_u16m4_m))) +void vlseg2e16(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_u16mf2_m))) +void vlseg2e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_u16mf4_m))) +void vlseg2e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_i16m1_m))) +void vlseg3e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_i16m2_m))) +void vlseg3e16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_i16mf2_m))) +void vlseg3e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_i16mf4_m))) +void vlseg3e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_u16m1_m))) +void vlseg3e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_u16m2_m))) +void vlseg3e16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_u16mf2_m))) +void vlseg3e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_u16mf4_m))) +void vlseg3e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_i16m1_m))) +void vlseg4e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_i16m2_m))) +void vlseg4e16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_i16mf2_m))) +void vlseg4e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_i16mf4_m))) +void vlseg4e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_u16m1_m))) +void vlseg4e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_u16m2_m))) +void vlseg4e16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_u16mf2_m))) +void vlseg4e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_u16mf4_m))) +void vlseg4e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_i16m1_m))) +void vlseg5e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_i16mf2_m))) +void vlseg5e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_i16mf4_m))) +void vlseg5e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_u16m1_m))) +void vlseg5e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_u16mf2_m))) +void vlseg5e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_u16mf4_m))) +void vlseg5e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_i16m1_m))) +void vlseg6e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_i16mf2_m))) +void vlseg6e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_i16mf4_m))) +void vlseg6e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_u16m1_m))) +void vlseg6e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_u16mf2_m))) +void vlseg6e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_u16mf4_m))) +void vlseg6e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_i16m1_m))) +void vlseg7e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_i16mf2_m))) +void vlseg7e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_i16mf4_m))) +void vlseg7e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_u16m1_m))) +void vlseg7e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_u16mf2_m))) +void vlseg7e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_u16mf4_m))) +void vlseg7e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_i16m1_m))) +void vlseg8e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_i16mf2_m))) +void vlseg8e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_i16mf4_m))) +void vlseg8e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_u16m1_m))) +void vlseg8e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_u16mf2_m))) +void vlseg8e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_u16mf4_m))) +void vlseg8e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_i32m1_m))) +void vlseg2e32(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_i32m2_m))) +void vlseg2e32(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_i32m4_m))) +void vlseg2e32(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_i32mf2_m))) +void vlseg2e32(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_u32m1_m))) +void vlseg2e32(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_u32m2_m))) +void vlseg2e32(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_u32m4_m))) +void vlseg2e32(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_u32mf2_m))) +void vlseg2e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_i32m1_m))) +void vlseg3e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_i32m2_m))) +void vlseg3e32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_i32mf2_m))) +void vlseg3e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_u32m1_m))) +void vlseg3e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_u32m2_m))) +void vlseg3e32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_u32mf2_m))) +void vlseg3e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_i32m1_m))) +void vlseg4e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_i32m2_m))) +void vlseg4e32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_i32mf2_m))) +void vlseg4e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_u32m1_m))) +void vlseg4e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_u32m2_m))) +void vlseg4e32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_u32mf2_m))) +void vlseg4e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32_v_i32m1_m))) +void vlseg5e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32_v_i32mf2_m))) +void vlseg5e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32_v_u32m1_m))) +void vlseg5e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32_v_u32mf2_m))) +void vlseg5e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32_v_i32m1_m))) +void vlseg6e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32_v_i32mf2_m))) +void vlseg6e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32_v_u32m1_m))) +void vlseg6e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32_v_u32mf2_m))) +void vlseg6e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32_v_i32m1_m))) +void vlseg7e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32_v_i32mf2_m))) +void vlseg7e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32_v_u32m1_m))) +void vlseg7e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32_v_u32mf2_m))) +void vlseg7e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32_v_i32m1_m))) +void vlseg8e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32_v_i32mf2_m))) +void vlseg8e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32_v_u32m1_m))) +void vlseg8e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32_v_u32mf2_m))) +void vlseg8e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_i64m1_m))) +void vlseg2e64(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_i64m2_m))) +void vlseg2e64(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_i64m4_m))) +void vlseg2e64(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_u64m1_m))) +void vlseg2e64(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_u64m2_m))) +void vlseg2e64(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_u64m4_m))) +void vlseg2e64(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64_v_i64m1_m))) +void vlseg3e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64_v_i64m2_m))) +void vlseg3e64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64_v_u64m1_m))) +void vlseg3e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64_v_u64m2_m))) +void vlseg3e64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64_v_i64m1_m))) +void vlseg4e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64_v_i64m2_m))) +void vlseg4e64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64_v_u64m1_m))) +void vlseg4e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64_v_u64m2_m))) +void vlseg4e64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e64_v_i64m1_m))) +void vlseg5e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e64_v_u64m1_m))) +void vlseg5e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e64_v_i64m1_m))) +void vlseg6e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e64_v_u64m1_m))) +void vlseg6e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e64_v_i64m1_m))) +void vlseg7e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e64_v_u64m1_m))) +void vlseg7e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e64_v_i64m1_m))) +void vlseg8e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e64_v_u64m1_m))) +void vlseg8e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_i8m1_m))) +void vlseg3e8ff(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_i8m2_m))) +void vlseg3e8ff(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_i8mf2_m))) +void vlseg3e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_i8mf4_m))) +void vlseg3e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_i8mf8_m))) +void vlseg3e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_u8m1_m))) +void vlseg3e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_u8m2_m))) +void vlseg3e8ff(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_u8mf2_m))) +void vlseg3e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_u8mf4_m))) +void vlseg3e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_u8mf8_m))) +void vlseg3e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_i8m1_m))) +void vlseg4e8ff(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_i8m2_m))) +void vlseg4e8ff(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_i8mf2_m))) +void vlseg4e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_i8mf4_m))) +void vlseg4e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_i8mf8_m))) +void vlseg4e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_u8m1_m))) +void vlseg4e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_u8m2_m))) +void vlseg4e8ff(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_u8mf2_m))) +void vlseg4e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_u8mf4_m))) +void vlseg4e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_u8mf8_m))) +void vlseg4e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_i8m1_m))) +void vlseg5e8ff(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_i8mf2_m))) +void vlseg5e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_i8mf4_m))) +void vlseg5e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_i8mf8_m))) +void vlseg5e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_u8m1_m))) +void vlseg5e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_u8mf2_m))) +void vlseg5e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_u8mf4_m))) +void vlseg5e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_u8mf8_m))) +void vlseg5e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_i8m1_m))) +void vlseg6e8ff(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_i8mf2_m))) +void vlseg6e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_i8mf4_m))) +void vlseg6e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_i8mf8_m))) +void vlseg6e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_u8m1_m))) +void vlseg6e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_u8mf2_m))) +void vlseg6e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_u8mf4_m))) +void vlseg6e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_u8mf8_m))) +void vlseg6e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_i8m1_m))) +void vlseg7e8ff(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_i8mf2_m))) +void vlseg7e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_i8mf4_m))) +void vlseg7e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_i8mf8_m))) +void vlseg7e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_u8m1_m))) +void vlseg7e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_u8mf2_m))) +void vlseg7e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_u8mf4_m))) +void vlseg7e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_u8mf8_m))) +void vlseg7e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_i8m1_m))) +void vlseg8e8ff(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_i8mf2_m))) +void vlseg8e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_i8mf4_m))) +void vlseg8e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_i8mf8_m))) +void vlseg8e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_u8m1_m))) +void vlseg8e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_u8mf2_m))) +void vlseg8e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_u8mf4_m))) +void vlseg8e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_u8mf8_m))) +void vlseg8e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_i16m1_m))) +void vlseg2e16ff(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_i16m2_m))) +void vlseg2e16ff(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_i16m4_m))) +void vlseg2e16ff(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_i16mf2_m))) +void vlseg2e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_i16mf4_m))) +void vlseg2e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_u16m1_m))) +void vlseg2e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_u16m2_m))) +void vlseg2e16ff(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_u16m4_m))) +void vlseg2e16ff(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_u16mf2_m))) +void vlseg2e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_u16mf4_m))) +void vlseg2e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_i16m1_m))) +void vlseg3e16ff(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_i16m2_m))) +void vlseg3e16ff(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_i16mf2_m))) +void vlseg3e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_i16mf4_m))) +void vlseg3e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_u16m1_m))) +void vlseg3e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_u16m2_m))) +void vlseg3e16ff(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_u16mf2_m))) +void vlseg3e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_u16mf4_m))) +void vlseg3e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_i16m1_m))) +void vlseg4e16ff(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_i16m2_m))) +void vlseg4e16ff(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_i16mf2_m))) +void vlseg4e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_i16mf4_m))) +void vlseg4e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_u16m1_m))) +void vlseg4e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_u16m2_m))) +void vlseg4e16ff(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_u16mf2_m))) +void vlseg4e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_u16mf4_m))) +void vlseg4e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_i16m1_m))) +void vlseg5e16ff(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_i16mf2_m))) +void vlseg5e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_i16mf4_m))) +void vlseg5e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_u16m1_m))) +void vlseg5e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_u16mf2_m))) +void vlseg5e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_u16mf4_m))) +void vlseg5e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_i16m1_m))) +void vlseg6e16ff(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_i16mf2_m))) +void vlseg6e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_i16mf4_m))) +void vlseg6e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_u16m1_m))) +void vlseg6e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_u16mf2_m))) +void vlseg6e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_u16mf4_m))) +void vlseg6e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_i16m1_m))) +void vlseg7e16ff(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_i16mf2_m))) +void vlseg7e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_i16mf4_m))) +void vlseg7e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_u16m1_m))) +void vlseg7e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_u16mf2_m))) +void vlseg7e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_u16mf4_m))) +void vlseg7e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_i16m1_m))) +void vlseg8e16ff(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_i16mf2_m))) +void vlseg8e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_i16mf4_m))) +void vlseg8e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_u16m1_m))) +void vlseg8e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_u16mf2_m))) +void vlseg8e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_u16mf4_m))) +void vlseg8e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_i32m1_m))) +void vlseg2e32ff(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_i32m2_m))) +void vlseg2e32ff(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_i32m4_m))) +void vlseg2e32ff(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_i32mf2_m))) +void vlseg2e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_i8m1_m))) +void vlseg2e8ff(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_i8m2_m))) +void vlseg2e8ff(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_i8m4_m))) +void vlseg2e8ff(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_i8mf2_m))) +void vlseg2e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_i8mf4_m))) +void vlseg2e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_i8mf8_m))) +void vlseg2e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_u32m1_m))) +void vlseg2e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_u32m2_m))) +void vlseg2e32ff(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_u32m4_m))) +void vlseg2e32ff(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_u32mf2_m))) +void vlseg2e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_i32m1_m))) +void vlseg3e32ff(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_i32m2_m))) +void vlseg3e32ff(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_i32mf2_m))) +void vlseg3e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_u32m1_m))) +void vlseg3e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_u32m2_m))) +void vlseg3e32ff(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_u32mf2_m))) +void vlseg3e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_i32m1_m))) +void vlseg4e32ff(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_i32m2_m))) +void vlseg4e32ff(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_i32mf2_m))) +void vlseg4e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_u32m1_m))) +void vlseg4e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_u32m2_m))) +void vlseg4e32ff(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_u32mf2_m))) +void vlseg4e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32ff_v_i32m1_m))) +void vlseg5e32ff(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32ff_v_i32mf2_m))) +void vlseg5e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32ff_v_u32m1_m))) +void vlseg5e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32ff_v_u32mf2_m))) +void vlseg5e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32ff_v_i32m1_m))) +void vlseg6e32ff(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32ff_v_i32mf2_m))) +void vlseg6e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32ff_v_u32m1_m))) +void vlseg6e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32ff_v_u32mf2_m))) +void vlseg6e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32ff_v_i32m1_m))) +void vlseg7e32ff(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32ff_v_i32mf2_m))) +void vlseg7e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_u8m1_m))) +void vlseg2e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_u8m2_m))) +void vlseg2e8ff(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_u8m4_m))) +void vlseg2e8ff(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_u8mf2_m))) +void vlseg2e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_u8mf4_m))) +void vlseg2e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_u8mf8_m))) +void vlseg2e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32ff_v_u32m1_m))) +void vlseg7e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32ff_v_u32mf2_m))) +void vlseg7e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32ff_v_i32m1_m))) +void vlseg8e32ff(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32ff_v_i32mf2_m))) +void vlseg8e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32ff_v_u32m1_m))) +void vlseg8e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32ff_v_u32mf2_m))) +void vlseg8e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_i64m1_m))) +void vlseg2e64ff(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_i64m2_m))) +void vlseg2e64ff(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_i64m4_m))) +void vlseg2e64ff(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_u64m1_m))) +void vlseg2e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_u64m2_m))) +void vlseg2e64ff(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_u64m4_m))) +void vlseg2e64ff(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64ff_v_i64m1_m))) +void vlseg3e64ff(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64ff_v_i64m2_m))) +void vlseg3e64ff(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64ff_v_u64m1_m))) +void vlseg3e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64ff_v_u64m2_m))) +void vlseg3e64ff(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64ff_v_i64m1_m))) +void vlseg4e64ff(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64ff_v_i64m2_m))) +void vlseg4e64ff(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64ff_v_u64m1_m))) +void vlseg4e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64ff_v_u64m2_m))) +void vlseg4e64ff(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e64ff_v_i64m1_m))) +void vlseg5e64ff(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_i8m1_m))) +void vlsseg2e8(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_i8m2_m))) +void vlsseg2e8(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_i8m4_m))) +void vlsseg2e8(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_i8mf2_m))) +void vlsseg2e8(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_i8mf4_m))) +void vlsseg2e8(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_i8mf8_m))) +void vlsseg2e8(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e64ff_v_u64m1_m))) +void vlseg5e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e64ff_v_i64m1_m))) +void vlseg6e64ff(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e64ff_v_u64m1_m))) +void vlseg6e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e64ff_v_i64m1_m))) +void vlseg7e64ff(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e64ff_v_u64m1_m))) +void vlseg7e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e64ff_v_i64m1_m))) +void vlseg8e64ff(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e64ff_v_u64m1_m))) +void vlseg8e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_u8m1_m))) +void vlsseg2e8(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_u8m2_m))) +void vlsseg2e8(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_u8m4_m))) +void vlsseg2e8(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_u8mf2_m))) +void vlsseg2e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_u8mf4_m))) +void vlsseg2e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_u8mf8_m))) +void vlsseg2e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8m1))) +void vluxseg2ei8(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8m1_m))) +void vluxseg2ei8(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8m2))) +void vluxseg2ei8(vint8m2_t * op0, vint8m2_t * op1, const int8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8m2_m))) +void vluxseg2ei8(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8m4))) +void vluxseg2ei8(vint8m4_t * op0, vint8m4_t * op1, const int8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8m4_m))) +void vluxseg2ei8(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, vuint8m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8mf2))) +void vluxseg2ei8(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8mf2_m))) +void vluxseg2ei8(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8mf4))) +void vluxseg2ei8(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8mf4_m))) +void vluxseg2ei8(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8mf8))) +void vluxseg2ei8(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8mf8_m))) +void vluxseg2ei8(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_i8m1_m))) +void vlsseg3e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_i8m2_m))) +void vlsseg3e8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_i8mf2_m))) +void vlsseg3e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_i8mf4_m))) +void vlsseg3e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_i8mf8_m))) +void vlsseg3e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8m1))) +void vluxseg2ei8(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8m1_m))) +void vluxseg2ei8(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8m2))) +void vluxseg2ei8(vuint8m2_t * op0, vuint8m2_t * op1, const uint8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8m2_m))) +void vluxseg2ei8(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8m4))) +void vluxseg2ei8(vuint8m4_t * op0, vuint8m4_t * op1, const uint8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8m4_m))) +void vluxseg2ei8(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, vuint8m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8mf2))) +void vluxseg2ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8mf2_m))) +void vluxseg2ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8mf4))) +void vluxseg2ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8mf4_m))) +void vluxseg2ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8mf8))) +void vluxseg2ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8mf8_m))) +void vluxseg2ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_u8m1_m))) +void vlsseg3e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_u8m2_m))) +void vlsseg3e8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_u8mf2_m))) +void vlsseg3e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_u8mf4_m))) +void vlsseg3e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_u8mf8_m))) +void vlsseg3e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_i8m1_m))) +void vlsseg4e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_i8m2_m))) +void vlsseg4e8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_i8mf2_m))) +void vlsseg4e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_i8mf4_m))) +void vlsseg4e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_i8mf8_m))) +void vlsseg4e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_u8m1_m))) +void vlsseg4e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_u8m2_m))) +void vlsseg4e8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_u8mf2_m))) +void vlsseg4e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_u8mf4_m))) +void vlsseg4e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_u8mf8_m))) +void vlsseg4e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_i8m1_m))) +void vlsseg5e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_i8mf2_m))) +void vlsseg5e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_i8mf4_m))) +void vlsseg5e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_i8mf8_m))) +void vlsseg5e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_u8m1_m))) +void vlsseg5e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_u8mf2_m))) +void vlsseg5e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_u8mf4_m))) +void vlsseg5e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_u8mf8_m))) +void vlsseg5e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_i8m1_m))) +void vlsseg6e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_i8mf2_m))) +void vlsseg6e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_i8mf4_m))) +void vlsseg6e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_i8mf8_m))) +void vlsseg6e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_u8m1_m))) +void vlsseg6e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_u8mf2_m))) +void vlsseg6e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_u8mf4_m))) +void vlsseg6e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_u8mf8_m))) +void vlsseg6e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_i8m1_m))) +void vlsseg7e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_i8mf2_m))) +void vlsseg7e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_i8mf4_m))) +void vlsseg7e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_i8mf8_m))) +void vlsseg7e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_u8m1_m))) +void vlsseg7e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_u8mf2_m))) +void vlsseg7e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_u8mf4_m))) +void vlsseg7e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_u8mf8_m))) +void vlsseg7e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_i8m1_m))) +void vlsseg8e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_i8mf2_m))) +void vlsseg8e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_i8mf4_m))) +void vlsseg8e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_i8mf8_m))) +void vlsseg8e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8m1))) +void vsseg2e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8m1_m))) +void vsseg2e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8m2))) +void vsseg2e8(int8_t * op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8m2_m))) +void vsseg2e8(vbool4_t op0, int8_t * op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8m4))) +void vsseg2e8(int8_t * op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8m4_m))) +void vsseg2e8(vbool2_t op0, int8_t * op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8mf2))) +void vsseg2e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8mf2_m))) +void vsseg2e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8mf4))) +void vsseg2e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8mf4_m))) +void vsseg2e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8mf8))) +void vsseg2e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8mf8_m))) +void vsseg2e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_u8m1_m))) +void vlsseg8e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_u8mf2_m))) +void vlsseg8e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_u8mf4_m))) +void vlsseg8e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_u8mf8_m))) +void vlsseg8e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_i16m1_m))) +void vlsseg2e16(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_i16m2_m))) +void vlsseg2e16(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_i16m4_m))) +void vlsseg2e16(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_i16mf2_m))) +void vlsseg2e16(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_i16mf4_m))) +void vlsseg2e16(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_u16m1_m))) +void vlsseg2e16(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_u16m2_m))) +void vlsseg2e16(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_u16m4_m))) +void vlsseg2e16(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_u16mf2_m))) +void vlsseg2e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_u16mf4_m))) +void vlsseg2e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_i16m1_m))) +void vlsseg3e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_i16m2_m))) +void vlsseg3e16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_i16mf2_m))) +void vlsseg3e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_i16mf4_m))) +void vlsseg3e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_u16m1_m))) +void vlsseg3e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_u16m2_m))) +void vlsseg3e16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_u16mf2_m))) +void vlsseg3e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_u16mf4_m))) +void vlsseg3e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_i16m1_m))) +void vlsseg4e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_i16m2_m))) +void vlsseg4e16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_i16mf2_m))) +void vlsseg4e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_i16mf4_m))) +void vlsseg4e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_u16m1_m))) +void vlsseg4e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_u16m2_m))) +void vlsseg4e16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_u16mf2_m))) +void vlsseg4e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_u16mf4_m))) +void vlsseg4e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_i16m1_m))) +void vlsseg5e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_i16mf2_m))) +void vlsseg5e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_i16mf4_m))) +void vlsseg5e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_u16m1_m))) +void vlsseg5e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_u16mf2_m))) +void vlsseg5e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_u16mf4_m))) +void vlsseg5e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_i16m1_m))) +void vlsseg6e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_i16mf2_m))) +void vlsseg6e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_i16mf4_m))) +void vlsseg6e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8m1))) +void vsseg2e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8m1_m))) +void vsseg2e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8m2))) +void vsseg2e8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8m2_m))) +void vsseg2e8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8m4))) +void vsseg2e8(uint8_t * op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8m4_m))) +void vsseg2e8(vbool2_t op0, uint8_t * op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8mf2))) +void vsseg2e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8mf2_m))) +void vsseg2e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8mf4))) +void vsseg2e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8mf4_m))) +void vsseg2e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8mf8))) +void vsseg2e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8mf8_m))) +void vsseg2e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_u16m1_m))) +void vlsseg6e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_u16mf2_m))) +void vlsseg6e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_u16mf4_m))) +void vlsseg6e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_i16m1_m))) +void vlsseg7e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_i16mf2_m))) +void vlsseg7e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_i16mf4_m))) +void vlsseg7e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_u16m1_m))) +void vlsseg7e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_u16mf2_m))) +void vlsseg7e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_u16mf4_m))) +void vlsseg7e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_i16m1_m))) +void vlsseg8e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_i16mf2_m))) +void vlsseg8e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_i16mf4_m))) +void vlsseg8e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_u16m1_m))) +void vlsseg8e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_u16mf2_m))) +void vlsseg8e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_u16mf4_m))) +void vlsseg8e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_i32m1_m))) +void vlsseg2e32(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_i32m2_m))) +void vlsseg2e32(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_i32m4_m))) +void vlsseg2e32(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_i32mf2_m))) +void vlsseg2e32(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_u32m1_m))) +void vlsseg2e32(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_u32m2_m))) +void vlsseg2e32(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_u32m4_m))) +void vlsseg2e32(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_u32mf2_m))) +void vlsseg2e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_i32m1_m))) +void vlsseg3e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_i32m2_m))) +void vlsseg3e32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_i32mf2_m))) +void vlsseg3e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_u32m1_m))) +void vlsseg3e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_u32m2_m))) +void vlsseg3e32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_u32mf2_m))) +void vlsseg3e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_i32m1_m))) +void vlsseg4e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_i32m2_m))) +void vlsseg4e32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_i32mf2_m))) +void vlsseg4e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_u32m1_m))) +void vlsseg4e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_u32m2_m))) +void vlsseg4e32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_u32mf2_m))) +void vlsseg4e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e32_v_i32m1_m))) +void vlsseg5e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e32_v_i32mf2_m))) +void vlsseg5e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e32_v_u32m1_m))) +void vlsseg5e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e32_v_u32mf2_m))) +void vlsseg5e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e32_v_i32m1_m))) +void vlsseg6e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e32_v_i32mf2_m))) +void vlsseg6e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e32_v_u32m1_m))) +void vlsseg6e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e32_v_u32mf2_m))) +void vlsseg6e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e32_v_i32m1_m))) +void vlsseg7e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e32_v_i32mf2_m))) +void vlsseg7e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e32_v_u32m1_m))) +void vlsseg7e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e32_v_u32mf2_m))) +void vlsseg7e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e32_v_i32m1_m))) +void vlsseg8e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e32_v_i32mf2_m))) +void vlsseg8e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e32_v_u32m1_m))) +void vlsseg8e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e32_v_u32mf2_m))) +void vlsseg8e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_i64m1_m))) +void vlsseg2e64(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_i64m2_m))) +void vlsseg2e64(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_i64m4_m))) +void vlsseg2e64(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_u64m1_m))) +void vlsseg2e64(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_u64m2_m))) +void vlsseg2e64(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_u64m4_m))) +void vlsseg2e64(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e64_v_i64m1_m))) +void vlsseg3e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e64_v_i64m2_m))) +void vlsseg3e64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e64_v_u64m1_m))) +void vlsseg3e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e64_v_u64m2_m))) +void vlsseg3e64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e64_v_i64m1_m))) +void vlsseg4e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e64_v_i64m2_m))) +void vlsseg4e64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e64_v_u64m1_m))) +void vlsseg4e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e64_v_u64m2_m))) +void vlsseg4e64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e64_v_i64m1_m))) +void vlsseg5e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e64_v_u64m1_m))) +void vlsseg5e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e64_v_i64m1_m))) +void vlsseg6e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e64_v_u64m1_m))) +void vlsseg6e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e64_v_i64m1_m))) +void vlsseg7e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e64_v_u64m1_m))) +void vlsseg7e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e64_v_i64m1_m))) +void vlsseg8e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e64_v_u64m1_m))) +void vlsseg8e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8m1))) +void vluxseg3ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8m1_m))) +void vluxseg3ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8m2))) +void vluxseg3ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, const int8_t * op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8m2_m))) +void vluxseg3ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, vuint8m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8mf2))) +void vluxseg3ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8mf2_m))) +void vluxseg3ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8mf4))) +void vluxseg3ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8mf4_m))) +void vluxseg3ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8mf8))) +void vluxseg3ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8mf8_m))) +void vluxseg3ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8m1))) +void vluxseg3ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8m1_m))) +void vluxseg3ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8m2))) +void vluxseg3ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, const uint8_t * op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8m2_m))) +void vluxseg3ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, vuint8m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8mf2))) +void vluxseg3ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8mf2_m))) +void vluxseg3ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8mf4))) +void vluxseg3ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8mf4_m))) +void vluxseg3ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8mf8))) +void vluxseg3ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8mf8_m))) +void vluxseg3ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8m1))) +void vluxseg4ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8m1_m))) +void vluxseg4ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8m2))) +void vluxseg4ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, const int8_t * op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8m2_m))) +void vluxseg4ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, vuint8m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8mf2))) +void vluxseg4ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8mf2_m))) +void vluxseg4ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8mf4))) +void vluxseg4ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8mf4_m))) +void vluxseg4ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8mf8))) +void vluxseg4ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8mf8_m))) +void vluxseg4ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8m1))) +void vluxseg4ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8m1_m))) +void vluxseg4ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8m2))) +void vluxseg4ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, const uint8_t * op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8m2_m))) +void vluxseg4ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, vuint8m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8mf2))) +void vluxseg4ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8mf2_m))) +void vluxseg4ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8mf4))) +void vluxseg4ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8mf4_m))) +void vluxseg4ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8mf8))) +void vluxseg4ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8mf8_m))) +void vluxseg4ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8m1))) +void vluxseg5ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8m1_m))) +void vluxseg5ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint8m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8mf2))) +void vluxseg5ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8mf2_m))) +void vluxseg5ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8mf4))) +void vluxseg5ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8mf4_m))) +void vluxseg5ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8mf8))) +void vluxseg5ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8mf8_m))) +void vluxseg5ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8m1))) +void vluxseg5ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8m1_m))) +void vluxseg5ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint8m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8mf2))) +void vluxseg5ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8mf2_m))) +void vluxseg5ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8mf4))) +void vluxseg5ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8mf4_m))) +void vluxseg5ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8mf8))) +void vluxseg5ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8mf8_m))) +void vluxseg5ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8m1))) +void vluxseg6ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8m1_m))) +void vluxseg6ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint8m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8mf2))) +void vluxseg6ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8mf2_m))) +void vluxseg6ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8mf4))) +void vluxseg6ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8mf4_m))) +void vluxseg6ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8mf8))) +void vluxseg6ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8mf8_m))) +void vluxseg6ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8m1))) +void vluxseg6ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8m1_m))) +void vluxseg6ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint8m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8mf2))) +void vluxseg6ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8mf2_m))) +void vluxseg6ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8mf4))) +void vluxseg6ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8mf4_m))) +void vluxseg6ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8mf8))) +void vluxseg6ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8mf8_m))) +void vluxseg6ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8m1))) +void vluxseg7ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8m1_m))) +void vluxseg7ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint8m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8mf2))) +void vluxseg7ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8mf2_m))) +void vluxseg7ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8mf4))) +void vluxseg7ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8mf4_m))) +void vluxseg7ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8mf8))) +void vluxseg7ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8mf8_m))) +void vluxseg7ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8m1))) +void vluxseg7ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8m1_m))) +void vluxseg7ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint8m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8mf2))) +void vluxseg7ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8mf2_m))) +void vluxseg7ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8mf4))) +void vluxseg7ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8mf4_m))) +void vluxseg7ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8mf8))) +void vluxseg7ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8mf8_m))) +void vluxseg7ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8m1))) +void vluxseg8ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8m1_m))) +void vluxseg8ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint8m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8mf2))) +void vluxseg8ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8mf2_m))) +void vluxseg8ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8mf4))) +void vluxseg8ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8mf4_m))) +void vluxseg8ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8mf8))) +void vluxseg8ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8mf8_m))) +void vluxseg8ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8m1))) +void vluxseg8ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8m1_m))) +void vluxseg8ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint8m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8mf2))) +void vluxseg8ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8mf2_m))) +void vluxseg8ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8mf4))) +void vluxseg8ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8mf4_m))) +void vluxseg8ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8mf8))) +void vluxseg8ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8mf8_m))) +void vluxseg8ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8m1))) +void vluxseg2ei16(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8m1_m))) +void vluxseg2ei16(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8m2))) +void vluxseg2ei16(vint8m2_t * op0, vint8m2_t * op1, const int8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8m2_m))) +void vluxseg2ei16(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8m4))) +void vluxseg2ei16(vint8m4_t * op0, vint8m4_t * op1, const int8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8m4_m))) +void vluxseg2ei16(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, vuint16m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8mf2))) +void vluxseg2ei16(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8mf2_m))) +void vluxseg2ei16(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8mf4))) +void vluxseg2ei16(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8mf4_m))) +void vluxseg2ei16(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8mf8))) +void vluxseg2ei16(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8mf8_m))) +void vluxseg2ei16(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8m1))) +void vluxseg2ei16(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8m1_m))) +void vluxseg2ei16(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8m2))) +void vluxseg2ei16(vuint8m2_t * op0, vuint8m2_t * op1, const uint8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8m2_m))) +void vluxseg2ei16(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8m4))) +void vluxseg2ei16(vuint8m4_t * op0, vuint8m4_t * op1, const uint8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8m4_m))) +void vluxseg2ei16(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, vuint16m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8mf2))) +void vluxseg2ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8mf2_m))) +void vluxseg2ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8mf4))) +void vluxseg2ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8mf4_m))) +void vluxseg2ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8mf8))) +void vluxseg2ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8mf8_m))) +void vluxseg2ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8m1))) +void vluxseg3ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8m1_m))) +void vluxseg3ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8m2))) +void vluxseg3ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, const int8_t * op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8m2_m))) +void vluxseg3ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, vuint16m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8mf2))) +void vluxseg3ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8mf2_m))) +void vluxseg3ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8mf4))) +void vluxseg3ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8mf4_m))) +void vluxseg3ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8mf8))) +void vluxseg3ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8mf8_m))) +void vluxseg3ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8m1))) +void vluxseg3ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8m1_m))) +void vluxseg3ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8m2))) +void vluxseg3ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, const uint8_t * op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8m2_m))) +void vluxseg3ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, vuint16m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8mf2))) +void vluxseg3ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8mf2_m))) +void vluxseg3ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8mf4))) +void vluxseg3ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8mf4_m))) +void vluxseg3ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8mf8))) +void vluxseg3ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8mf8_m))) +void vluxseg3ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8m1))) +void vluxseg4ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8m1_m))) +void vluxseg4ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8m2))) +void vluxseg4ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, const int8_t * op4, vuint16m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8m2_m))) +void vluxseg4ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, vuint16m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8mf2))) +void vluxseg4ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8mf2_m))) +void vluxseg4ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8mf4))) +void vluxseg4ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8mf4_m))) +void vluxseg4ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8mf8))) +void vluxseg4ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8mf8_m))) +void vluxseg4ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8m1))) +void vluxseg4ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8m1_m))) +void vluxseg4ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8m2))) +void vluxseg4ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, const uint8_t * op4, vuint16m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8m2_m))) +void vluxseg4ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, vuint16m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8mf2))) +void vluxseg4ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8mf2_m))) +void vluxseg4ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8mf4))) +void vluxseg4ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8mf4_m))) +void vluxseg4ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8mf8))) +void vluxseg4ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8mf8_m))) +void vluxseg4ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8m1))) +void vluxseg5ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8m1_m))) +void vluxseg5ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint16m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8mf2))) +void vluxseg5ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8mf2_m))) +void vluxseg5ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8mf4))) +void vluxseg5ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8mf4_m))) +void vluxseg5ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8mf8))) +void vluxseg5ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8mf8_m))) +void vluxseg5ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8m1))) +void vluxseg5ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8m1_m))) +void vluxseg5ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint16m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8mf2))) +void vluxseg5ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8mf2_m))) +void vluxseg5ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8mf4))) +void vluxseg5ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8mf4_m))) +void vluxseg5ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8mf8))) +void vluxseg5ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8mf8_m))) +void vluxseg5ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8m1))) +void vluxseg6ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint16m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8m1_m))) +void vluxseg6ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint16m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8mf2))) +void vluxseg6ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8mf2_m))) +void vluxseg6ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8mf4))) +void vluxseg6ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8mf4_m))) +void vluxseg6ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8mf8))) +void vluxseg6ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8mf8_m))) +void vluxseg6ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8m1))) +void vluxseg6ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint16m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8m1_m))) +void vluxseg6ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint16m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8mf2))) +void vluxseg6ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8mf2_m))) +void vluxseg6ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8mf4))) +void vluxseg6ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8mf4_m))) +void vluxseg6ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8mf8))) +void vluxseg6ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8mf8_m))) +void vluxseg6ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8m1))) +void vluxseg7ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8m1_m))) +void vluxseg7ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint16m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8mf2))) +void vluxseg7ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8mf2_m))) +void vluxseg7ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8mf4))) +void vluxseg7ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8mf4_m))) +void vluxseg7ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8mf8))) +void vluxseg7ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8mf8_m))) +void vluxseg7ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8m1))) +void vluxseg7ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8m1_m))) +void vluxseg7ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint16m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8mf2))) +void vluxseg7ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8mf2_m))) +void vluxseg7ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8mf4))) +void vluxseg7ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8mf4_m))) +void vluxseg7ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8mf8))) +void vluxseg7ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8mf8_m))) +void vluxseg7ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8m1))) +void vluxseg8ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint16m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8m1_m))) +void vluxseg8ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint16m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8mf2))) +void vluxseg8ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8mf2_m))) +void vluxseg8ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8mf4))) +void vluxseg8ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8mf4_m))) +void vluxseg8ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8mf8))) +void vluxseg8ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8mf8_m))) +void vluxseg8ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8m1))) +void vluxseg8ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint16m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8m1_m))) +void vluxseg8ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint16m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8mf2))) +void vluxseg8ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8mf2_m))) +void vluxseg8ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8mf4))) +void vluxseg8ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8mf4_m))) +void vluxseg8ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8mf8))) +void vluxseg8ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8mf8_m))) +void vluxseg8ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8m1))) +void vluxseg2ei32(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8m1_m))) +void vluxseg2ei32(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8m2))) +void vluxseg2ei32(vint8m2_t * op0, vint8m2_t * op1, const int8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8m2_m))) +void vluxseg2ei32(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8mf2))) +void vluxseg2ei32(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8mf2_m))) +void vluxseg2ei32(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8mf4))) +void vluxseg2ei32(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8mf4_m))) +void vluxseg2ei32(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8mf8))) +void vluxseg2ei32(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8mf8_m))) +void vluxseg2ei32(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8m1))) +void vluxseg2ei32(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8m1_m))) +void vluxseg2ei32(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8m2))) +void vluxseg2ei32(vuint8m2_t * op0, vuint8m2_t * op1, const uint8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8m2_m))) +void vluxseg2ei32(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8mf2))) +void vluxseg2ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8mf2_m))) +void vluxseg2ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8mf4))) +void vluxseg2ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8mf4_m))) +void vluxseg2ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8mf8))) +void vluxseg2ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8mf8_m))) +void vluxseg2ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8m1))) +void vluxseg3ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8m1_m))) +void vluxseg3ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8m2))) +void vluxseg3ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, const int8_t * op3, vuint32m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8m2_m))) +void vluxseg3ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, vuint32m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8mf2))) +void vluxseg3ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8mf2_m))) +void vluxseg3ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8mf4))) +void vluxseg3ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8mf4_m))) +void vluxseg3ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8mf8))) +void vluxseg3ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8mf8_m))) +void vluxseg3ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8m1))) +void vluxseg3ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8m1_m))) +void vluxseg3ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8m2))) +void vluxseg3ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, const uint8_t * op3, vuint32m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8m2_m))) +void vluxseg3ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, vuint32m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8mf2))) +void vluxseg3ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8mf2_m))) +void vluxseg3ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8mf4))) +void vluxseg3ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8mf4_m))) +void vluxseg3ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8mf8))) +void vluxseg3ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8mf8_m))) +void vluxseg3ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8m1))) +void vluxseg4ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8m1_m))) +void vluxseg4ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8m2))) +void vluxseg4ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, const int8_t * op4, vuint32m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8m2_m))) +void vluxseg4ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, vuint32m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8mf2))) +void vluxseg4ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8mf2_m))) +void vluxseg4ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8mf4))) +void vluxseg4ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8mf4_m))) +void vluxseg4ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8mf8))) +void vluxseg4ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8mf8_m))) +void vluxseg4ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8m1))) +void vluxseg4ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8m1_m))) +void vluxseg4ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8m2))) +void vluxseg4ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, const uint8_t * op4, vuint32m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8m2_m))) +void vluxseg4ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, vuint32m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8mf2))) +void vluxseg4ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8mf2_m))) +void vluxseg4ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8mf4))) +void vluxseg4ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8mf4_m))) +void vluxseg4ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8mf8))) +void vluxseg4ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8mf8_m))) +void vluxseg4ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8m1))) +void vluxseg5ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8m1_m))) +void vluxseg5ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint32m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8mf2))) +void vluxseg5ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8mf2_m))) +void vluxseg5ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8mf4))) +void vluxseg5ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8mf4_m))) +void vluxseg5ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8mf8))) +void vluxseg5ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8mf8_m))) +void vluxseg5ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8m1))) +void vluxseg5ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8m1_m))) +void vluxseg5ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint32m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8mf2))) +void vluxseg5ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8mf2_m))) +void vluxseg5ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8mf4))) +void vluxseg5ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8mf4_m))) +void vluxseg5ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8mf8))) +void vluxseg5ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8mf8_m))) +void vluxseg5ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8m1))) +void vluxseg6ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint32m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8m1_m))) +void vluxseg6ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint32m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8mf2))) +void vluxseg6ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8mf2_m))) +void vluxseg6ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8mf4))) +void vluxseg6ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8mf4_m))) +void vluxseg6ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8mf8))) +void vluxseg6ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8mf8_m))) +void vluxseg6ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8m1))) +void vluxseg6ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint32m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8m1_m))) +void vluxseg6ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint32m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8mf2))) +void vluxseg6ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8mf2_m))) +void vluxseg6ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8mf4))) +void vluxseg6ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8mf4_m))) +void vluxseg6ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8mf8))) +void vluxseg6ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8mf8_m))) +void vluxseg6ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8m1))) +void vluxseg7ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8m1_m))) +void vluxseg7ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint32m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8mf2))) +void vluxseg7ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8mf2_m))) +void vluxseg7ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8mf4))) +void vluxseg7ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8mf4_m))) +void vluxseg7ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8mf8))) +void vluxseg7ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8mf8_m))) +void vluxseg7ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8m1))) +void vluxseg7ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8m1_m))) +void vluxseg7ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint32m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8mf2))) +void vluxseg7ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8mf2_m))) +void vluxseg7ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8mf4))) +void vluxseg7ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8mf4_m))) +void vluxseg7ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8mf8))) +void vluxseg7ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8mf8_m))) +void vluxseg7ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8m1))) +void vluxseg8ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint32m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8m1_m))) +void vluxseg8ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint32m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8mf2))) +void vluxseg8ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8mf2_m))) +void vluxseg8ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8mf4))) +void vluxseg8ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8mf4_m))) +void vluxseg8ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8mf8))) +void vluxseg8ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8mf8_m))) +void vluxseg8ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8m1))) +void vluxseg8ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint32m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8m1_m))) +void vluxseg8ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint32m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8mf2))) +void vluxseg8ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8mf2_m))) +void vluxseg8ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8mf4))) +void vluxseg8ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8mf4_m))) +void vluxseg8ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8mf8))) +void vluxseg8ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8mf8_m))) +void vluxseg8ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8m1))) +void vluxseg2ei64(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8m1_m))) +void vluxseg2ei64(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8mf2))) +void vluxseg2ei64(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8mf2_m))) +void vluxseg2ei64(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8mf4))) +void vluxseg2ei64(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8mf4_m))) +void vluxseg2ei64(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8mf8))) +void vluxseg2ei64(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8mf8_m))) +void vluxseg2ei64(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8m1))) +void vluxseg2ei64(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8m1_m))) +void vluxseg2ei64(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8mf2))) +void vluxseg2ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8mf2_m))) +void vluxseg2ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8mf4))) +void vluxseg2ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8mf4_m))) +void vluxseg2ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8mf8))) +void vluxseg2ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8mf8_m))) +void vluxseg2ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8m1))) +void vluxseg3ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8m1_m))) +void vluxseg3ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8mf2))) +void vluxseg3ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8mf2_m))) +void vluxseg3ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8mf4))) +void vluxseg3ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8mf4_m))) +void vluxseg3ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8mf8))) +void vluxseg3ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8mf8_m))) +void vluxseg3ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8m1))) +void vluxseg3ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8m1_m))) +void vluxseg3ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8mf2))) +void vluxseg3ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8mf2_m))) +void vluxseg3ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8mf4))) +void vluxseg3ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8mf4_m))) +void vluxseg3ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8mf8))) +void vluxseg3ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8mf8_m))) +void vluxseg3ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8m1))) +void vluxseg4ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8m1_m))) +void vluxseg4ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8mf2))) +void vluxseg4ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8mf2_m))) +void vluxseg4ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8mf4))) +void vluxseg4ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8mf4_m))) +void vluxseg4ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8mf8))) +void vluxseg4ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8mf8_m))) +void vluxseg4ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8m1))) +void vluxseg4ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8m1_m))) +void vluxseg4ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8mf2))) +void vluxseg4ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8mf2_m))) +void vluxseg4ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8mf4))) +void vluxseg4ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8mf4_m))) +void vluxseg4ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8mf8))) +void vluxseg4ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8mf8_m))) +void vluxseg4ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8m1))) +void vluxseg5ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8m1_m))) +void vluxseg5ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint64m8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8mf2))) +void vluxseg5ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8mf2_m))) +void vluxseg5ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8mf4))) +void vluxseg5ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8mf4_m))) +void vluxseg5ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8mf8))) +void vluxseg5ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8mf8_m))) +void vluxseg5ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8m1))) +void vluxseg5ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8m1_m))) +void vluxseg5ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint64m8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8mf2))) +void vluxseg5ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8mf2_m))) +void vluxseg5ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8mf4))) +void vluxseg5ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8mf4_m))) +void vluxseg5ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8mf8))) +void vluxseg5ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8mf8_m))) +void vluxseg5ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8m1))) +void vluxseg6ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint64m8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8m1_m))) +void vluxseg6ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint64m8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8mf2))) +void vluxseg6ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8mf2_m))) +void vluxseg6ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8mf4))) +void vluxseg6ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8mf4_m))) +void vluxseg6ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8mf8))) +void vluxseg6ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8mf8_m))) +void vluxseg6ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8m1))) +void vluxseg6ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint64m8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8m1_m))) +void vluxseg6ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint64m8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8mf2))) +void vluxseg6ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8mf2_m))) +void vluxseg6ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8mf4))) +void vluxseg6ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8mf4_m))) +void vluxseg6ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8mf8))) +void vluxseg6ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8mf8_m))) +void vluxseg6ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8m1))) +void vluxseg7ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8m1_m))) +void vluxseg7ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint64m8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8mf2))) +void vluxseg7ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8mf2_m))) +void vluxseg7ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8mf4))) +void vluxseg7ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8mf4_m))) +void vluxseg7ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8mf8))) +void vluxseg7ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8mf8_m))) +void vluxseg7ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8m1))) +void vluxseg7ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8m1_m))) +void vluxseg7ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint64m8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8mf2))) +void vluxseg7ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8mf2_m))) +void vluxseg7ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8mf4))) +void vluxseg7ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8mf4_m))) +void vluxseg7ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8mf8))) +void vluxseg7ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8mf8_m))) +void vluxseg7ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8m1))) +void vluxseg8ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint64m8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8m1_m))) +void vluxseg8ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint64m8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8mf2))) +void vluxseg8ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8mf2_m))) +void vluxseg8ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8mf4))) +void vluxseg8ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8mf4_m))) +void vluxseg8ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8mf8))) +void vluxseg8ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8mf8_m))) +void vluxseg8ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8m1))) +void vluxseg8ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint64m8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8m1_m))) +void vluxseg8ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint64m8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8mf2))) +void vluxseg8ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8mf2_m))) +void vluxseg8ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8mf4))) +void vluxseg8ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8mf4_m))) +void vluxseg8ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8mf8))) +void vluxseg8ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8mf8_m))) +void vluxseg8ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16m1))) +void vluxseg2ei8(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16m1_m))) +void vluxseg2ei8(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16m2))) +void vluxseg2ei8(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16m2_m))) +void vluxseg2ei8(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16m4))) +void vluxseg2ei8(vint16m4_t * op0, vint16m4_t * op1, const int16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16m4_m))) +void vluxseg2ei8(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16mf2))) +void vluxseg2ei8(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16mf2_m))) +void vluxseg2ei8(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16mf4))) +void vluxseg2ei8(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16mf4_m))) +void vluxseg2ei8(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16m1))) +void vluxseg2ei8(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16m1_m))) +void vluxseg2ei8(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16m2))) +void vluxseg2ei8(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16m2_m))) +void vluxseg2ei8(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16m4))) +void vluxseg2ei8(vuint16m4_t * op0, vuint16m4_t * op1, const uint16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16m4_m))) +void vluxseg2ei8(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16mf2))) +void vluxseg2ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16mf2_m))) +void vluxseg2ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16mf4))) +void vluxseg2ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16mf4_m))) +void vluxseg2ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16m1))) +void vluxseg3ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16m1_m))) +void vluxseg3ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16m2))) +void vluxseg3ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16m2_m))) +void vluxseg3ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16mf2))) +void vluxseg3ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16mf2_m))) +void vluxseg3ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16mf4))) +void vluxseg3ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16mf4_m))) +void vluxseg3ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16m1))) +void vluxseg3ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16m1_m))) +void vluxseg3ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16m2))) +void vluxseg3ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16m2_m))) +void vluxseg3ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16mf2))) +void vluxseg3ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16mf2_m))) +void vluxseg3ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16mf4))) +void vluxseg3ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16mf4_m))) +void vluxseg3ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16m1))) +void vluxseg4ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16m1_m))) +void vluxseg4ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16m2))) +void vluxseg4ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16m2_m))) +void vluxseg4ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16mf2))) +void vluxseg4ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16mf2_m))) +void vluxseg4ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16mf4))) +void vluxseg4ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16mf4_m))) +void vluxseg4ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16m1))) +void vluxseg4ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16m1_m))) +void vluxseg4ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16m2))) +void vluxseg4ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16m2_m))) +void vluxseg4ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16mf2))) +void vluxseg4ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16mf2_m))) +void vluxseg4ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16mf4))) +void vluxseg4ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16mf4_m))) +void vluxseg4ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i16m1))) +void vluxseg5ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i16m1_m))) +void vluxseg5ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i16mf2))) +void vluxseg5ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i16mf2_m))) +void vluxseg5ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i16mf4))) +void vluxseg5ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i16mf4_m))) +void vluxseg5ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u16m1))) +void vluxseg5ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u16m1_m))) +void vluxseg5ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u16mf2))) +void vluxseg5ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u16mf2_m))) +void vluxseg5ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u16mf4))) +void vluxseg5ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u16mf4_m))) +void vluxseg5ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i16m1))) +void vluxseg6ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i16m1_m))) +void vluxseg6ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i16mf2))) +void vluxseg6ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i16mf2_m))) +void vluxseg6ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i16mf4))) +void vluxseg6ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i16mf4_m))) +void vluxseg6ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u16m1))) +void vluxseg6ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u16m1_m))) +void vluxseg6ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u16mf2))) +void vluxseg6ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u16mf2_m))) +void vluxseg6ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u16mf4))) +void vluxseg6ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u16mf4_m))) +void vluxseg6ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i16m1))) +void vluxseg7ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i16m1_m))) +void vluxseg7ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i16mf2))) +void vluxseg7ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i16mf2_m))) +void vluxseg7ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i16mf4))) +void vluxseg7ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i16mf4_m))) +void vluxseg7ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u16m1))) +void vluxseg7ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u16m1_m))) +void vluxseg7ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u16mf2))) +void vluxseg7ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u16mf2_m))) +void vluxseg7ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u16mf4))) +void vluxseg7ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u16mf4_m))) +void vluxseg7ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i16m1))) +void vluxseg8ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i16m1_m))) +void vluxseg8ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i16mf2))) +void vluxseg8ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i16mf2_m))) +void vluxseg8ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i16mf4))) +void vluxseg8ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i16mf4_m))) +void vluxseg8ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u16m1))) +void vluxseg8ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u16m1_m))) +void vluxseg8ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u16mf2))) +void vluxseg8ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u16mf2_m))) +void vluxseg8ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u16mf4))) +void vluxseg8ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u16mf4_m))) +void vluxseg8ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16m1))) +void vluxseg2ei16(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16m1_m))) +void vluxseg2ei16(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16m2))) +void vluxseg2ei16(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16m2_m))) +void vluxseg2ei16(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16m4))) +void vluxseg2ei16(vint16m4_t * op0, vint16m4_t * op1, const int16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16m4_m))) +void vluxseg2ei16(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16mf2))) +void vluxseg2ei16(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16mf2_m))) +void vluxseg2ei16(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16mf4))) +void vluxseg2ei16(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16mf4_m))) +void vluxseg2ei16(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16m1))) +void vluxseg2ei16(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16m1_m))) +void vluxseg2ei16(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16m2))) +void vluxseg2ei16(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16m2_m))) +void vluxseg2ei16(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16m4))) +void vluxseg2ei16(vuint16m4_t * op0, vuint16m4_t * op1, const uint16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16m4_m))) +void vluxseg2ei16(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16mf2))) +void vluxseg2ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16mf2_m))) +void vluxseg2ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16mf4))) +void vluxseg2ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16mf4_m))) +void vluxseg2ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16m1))) +void vluxseg3ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16m1_m))) +void vluxseg3ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16m2))) +void vluxseg3ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16m2_m))) +void vluxseg3ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16mf2))) +void vluxseg3ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16mf2_m))) +void vluxseg3ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16mf4))) +void vluxseg3ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16mf4_m))) +void vluxseg3ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16m1))) +void vluxseg3ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16m1_m))) +void vluxseg3ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16m2))) +void vluxseg3ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16m2_m))) +void vluxseg3ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16mf2))) +void vluxseg3ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16mf2_m))) +void vluxseg3ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16mf4))) +void vluxseg3ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16mf4_m))) +void vluxseg3ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16m1))) +void vluxseg4ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16m1_m))) +void vluxseg4ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16m2))) +void vluxseg4ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16m2_m))) +void vluxseg4ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16mf2))) +void vluxseg4ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16mf2_m))) +void vluxseg4ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16mf4))) +void vluxseg4ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16mf4_m))) +void vluxseg4ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16m1))) +void vluxseg4ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16m1_m))) +void vluxseg4ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16m2))) +void vluxseg4ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16m2_m))) +void vluxseg4ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16mf2))) +void vluxseg4ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16mf2_m))) +void vluxseg4ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16mf4))) +void vluxseg4ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16mf4_m))) +void vluxseg4ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i16m1))) +void vluxseg5ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i16m1_m))) +void vluxseg5ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i16mf2))) +void vluxseg5ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i16mf2_m))) +void vluxseg5ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i16mf4))) +void vluxseg5ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i16mf4_m))) +void vluxseg5ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u16m1))) +void vluxseg5ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u16m1_m))) +void vluxseg5ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u16mf2))) +void vluxseg5ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u16mf2_m))) +void vluxseg5ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u16mf4))) +void vluxseg5ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u16mf4_m))) +void vluxseg5ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i16m1))) +void vluxseg6ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i16m1_m))) +void vluxseg6ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i16mf2))) +void vluxseg6ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i16mf2_m))) +void vluxseg6ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i16mf4))) +void vluxseg6ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i16mf4_m))) +void vluxseg6ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u16m1))) +void vluxseg6ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u16m1_m))) +void vluxseg6ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u16mf2))) +void vluxseg6ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u16mf2_m))) +void vluxseg6ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u16mf4))) +void vluxseg6ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u16mf4_m))) +void vluxseg6ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i16m1))) +void vluxseg7ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i16m1_m))) +void vluxseg7ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i16mf2))) +void vluxseg7ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i16mf2_m))) +void vluxseg7ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i16mf4))) +void vluxseg7ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i16mf4_m))) +void vluxseg7ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u16m1))) +void vluxseg7ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u16m1_m))) +void vluxseg7ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u16mf2))) +void vluxseg7ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u16mf2_m))) +void vluxseg7ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u16mf4))) +void vluxseg7ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u16mf4_m))) +void vluxseg7ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i16m1))) +void vluxseg8ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i16m1_m))) +void vluxseg8ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i16mf2))) +void vluxseg8ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i16mf2_m))) +void vluxseg8ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i16mf4))) +void vluxseg8ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i16mf4_m))) +void vluxseg8ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u16m1))) +void vluxseg8ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u16m1_m))) +void vluxseg8ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u16mf2))) +void vluxseg8ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u16mf2_m))) +void vluxseg8ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u16mf4))) +void vluxseg8ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u16mf4_m))) +void vluxseg8ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16m1))) +void vluxseg2ei32(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16m1_m))) +void vluxseg2ei32(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16m2))) +void vluxseg2ei32(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16m2_m))) +void vluxseg2ei32(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16m4))) +void vluxseg2ei32(vint16m4_t * op0, vint16m4_t * op1, const int16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16m4_m))) +void vluxseg2ei32(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16mf2))) +void vluxseg2ei32(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16mf2_m))) +void vluxseg2ei32(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16mf4))) +void vluxseg2ei32(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16mf4_m))) +void vluxseg2ei32(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16m1))) +void vluxseg2ei32(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16m1_m))) +void vluxseg2ei32(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16m2))) +void vluxseg2ei32(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16m2_m))) +void vluxseg2ei32(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16m4))) +void vluxseg2ei32(vuint16m4_t * op0, vuint16m4_t * op1, const uint16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16m4_m))) +void vluxseg2ei32(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16mf2))) +void vluxseg2ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16mf2_m))) +void vluxseg2ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16mf4))) +void vluxseg2ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16mf4_m))) +void vluxseg2ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16m1))) +void vluxseg3ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16m1_m))) +void vluxseg3ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16m2))) +void vluxseg3ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16m2_m))) +void vluxseg3ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16mf2))) +void vluxseg3ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16mf2_m))) +void vluxseg3ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16mf4))) +void vluxseg3ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16mf4_m))) +void vluxseg3ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16m1))) +void vluxseg3ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16m1_m))) +void vluxseg3ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16m2))) +void vluxseg3ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16m2_m))) +void vluxseg3ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16mf2))) +void vluxseg3ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16mf2_m))) +void vluxseg3ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16mf4))) +void vluxseg3ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16mf4_m))) +void vluxseg3ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16m1))) +void vluxseg4ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16m1_m))) +void vluxseg4ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16m2))) +void vluxseg4ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16m2_m))) +void vluxseg4ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16mf2))) +void vluxseg4ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16mf2_m))) +void vluxseg4ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16mf4))) +void vluxseg4ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16mf4_m))) +void vluxseg4ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16m1))) +void vluxseg4ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16m1_m))) +void vluxseg4ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16m2))) +void vluxseg4ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16m2_m))) +void vluxseg4ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16mf2))) +void vluxseg4ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16mf2_m))) +void vluxseg4ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16mf4))) +void vluxseg4ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16mf4_m))) +void vluxseg4ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i16m1))) +void vluxseg5ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i16m1_m))) +void vluxseg5ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i16mf2))) +void vluxseg5ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i16mf2_m))) +void vluxseg5ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i16mf4))) +void vluxseg5ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i16mf4_m))) +void vluxseg5ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u16m1))) +void vluxseg5ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u16m1_m))) +void vluxseg5ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u16mf2))) +void vluxseg5ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u16mf2_m))) +void vluxseg5ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u16mf4))) +void vluxseg5ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u16mf4_m))) +void vluxseg5ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i16m1))) +void vluxseg6ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i16m1_m))) +void vluxseg6ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i16mf2))) +void vluxseg6ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i16mf2_m))) +void vluxseg6ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i16mf4))) +void vluxseg6ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i16mf4_m))) +void vluxseg6ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u16m1))) +void vluxseg6ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u16m1_m))) +void vluxseg6ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u16mf2))) +void vluxseg6ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u16mf2_m))) +void vluxseg6ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u16mf4))) +void vluxseg6ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u16mf4_m))) +void vluxseg6ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i16m1))) +void vluxseg7ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i16m1_m))) +void vluxseg7ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i16mf2))) +void vluxseg7ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i16mf2_m))) +void vluxseg7ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i16mf4))) +void vluxseg7ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i16mf4_m))) +void vluxseg7ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u16m1))) +void vluxseg7ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u16m1_m))) +void vluxseg7ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u16mf2))) +void vluxseg7ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u16mf2_m))) +void vluxseg7ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u16mf4))) +void vluxseg7ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u16mf4_m))) +void vluxseg7ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i16m1))) +void vluxseg8ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i16m1_m))) +void vluxseg8ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i16mf2))) +void vluxseg8ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i16mf2_m))) +void vluxseg8ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i16mf4))) +void vluxseg8ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i16mf4_m))) +void vluxseg8ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8m1))) +void vssseg2e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8m1_m))) +void vssseg2e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8m2))) +void vssseg2e8(int8_t * op0, ptrdiff_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8m2_m))) +void vssseg2e8(vbool4_t op0, int8_t * op1, ptrdiff_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8m4))) +void vssseg2e8(int8_t * op0, ptrdiff_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8m4_m))) +void vssseg2e8(vbool2_t op0, int8_t * op1, ptrdiff_t op2, vint8m4_t op3, vint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8mf2))) +void vssseg2e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8mf2_m))) +void vssseg2e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8mf4))) +void vssseg2e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8mf4_m))) +void vssseg2e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8mf8))) +void vssseg2e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8mf8_m))) +void vssseg2e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u16m1))) +void vluxseg8ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u16m1_m))) +void vluxseg8ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u16mf2))) +void vluxseg8ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u16mf2_m))) +void vluxseg8ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u16mf4))) +void vluxseg8ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u16mf4_m))) +void vluxseg8ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16m1))) +void vluxseg2ei64(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16m1_m))) +void vluxseg2ei64(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16m2))) +void vluxseg2ei64(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16m2_m))) +void vluxseg2ei64(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16mf2))) +void vluxseg2ei64(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16mf2_m))) +void vluxseg2ei64(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16mf4))) +void vluxseg2ei64(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16mf4_m))) +void vluxseg2ei64(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16m1))) +void vluxseg2ei64(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16m1_m))) +void vluxseg2ei64(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16m2))) +void vluxseg2ei64(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16m2_m))) +void vluxseg2ei64(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16mf2))) +void vluxseg2ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16mf2_m))) +void vluxseg2ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16mf4))) +void vluxseg2ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16mf4_m))) +void vluxseg2ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16m1))) +void vluxseg3ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16m1_m))) +void vluxseg3ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16m2))) +void vluxseg3ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16m2_m))) +void vluxseg3ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16mf2))) +void vluxseg3ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16mf2_m))) +void vluxseg3ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16mf4))) +void vluxseg3ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16mf4_m))) +void vluxseg3ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16m1))) +void vluxseg3ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16m1_m))) +void vluxseg3ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16m2))) +void vluxseg3ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16m2_m))) +void vluxseg3ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16mf2))) +void vluxseg3ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16mf2_m))) +void vluxseg3ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16mf4))) +void vluxseg3ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16mf4_m))) +void vluxseg3ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16m1))) +void vluxseg4ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16m1_m))) +void vluxseg4ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16m2))) +void vluxseg4ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16m2_m))) +void vluxseg4ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16mf2))) +void vluxseg4ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16mf2_m))) +void vluxseg4ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16mf4))) +void vluxseg4ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16mf4_m))) +void vluxseg4ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16m1))) +void vluxseg4ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16m1_m))) +void vluxseg4ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16m2))) +void vluxseg4ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16m2_m))) +void vluxseg4ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16mf2))) +void vluxseg4ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16mf2_m))) +void vluxseg4ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16mf4))) +void vluxseg4ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16mf4_m))) +void vluxseg4ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i16m1))) +void vluxseg5ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i16m1_m))) +void vluxseg5ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i16mf2))) +void vluxseg5ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i16mf2_m))) +void vluxseg5ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i16mf4))) +void vluxseg5ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i16mf4_m))) +void vluxseg5ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u16m1))) +void vluxseg5ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u16m1_m))) +void vluxseg5ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u16mf2))) +void vluxseg5ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u16mf2_m))) +void vluxseg5ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u16mf4))) +void vluxseg5ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u16mf4_m))) +void vluxseg5ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i16m1))) +void vluxseg6ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i16m1_m))) +void vluxseg6ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i16mf2))) +void vluxseg6ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i16mf2_m))) +void vluxseg6ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i16mf4))) +void vluxseg6ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i16mf4_m))) +void vluxseg6ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8m1))) +void vssseg2e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8m1_m))) +void vssseg2e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8m2))) +void vssseg2e8(uint8_t * op0, ptrdiff_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8m2_m))) +void vssseg2e8(vbool4_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8m4))) +void vssseg2e8(uint8_t * op0, ptrdiff_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8m4_m))) +void vssseg2e8(vbool2_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m4_t op3, vuint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8mf2))) +void vssseg2e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8mf2_m))) +void vssseg2e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8mf4))) +void vssseg2e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8mf4_m))) +void vssseg2e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8mf8))) +void vssseg2e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8mf8_m))) +void vssseg2e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u16m1))) +void vluxseg6ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u16m1_m))) +void vluxseg6ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u16mf2))) +void vluxseg6ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u16mf2_m))) +void vluxseg6ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u16mf4))) +void vluxseg6ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u16mf4_m))) +void vluxseg6ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i16m1))) +void vluxseg7ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i16m1_m))) +void vluxseg7ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i16mf2))) +void vluxseg7ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i16mf2_m))) +void vluxseg7ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i16mf4))) +void vluxseg7ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i16mf4_m))) +void vluxseg7ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u16m1))) +void vluxseg7ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u16m1_m))) +void vluxseg7ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u16mf2))) +void vluxseg7ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u16mf2_m))) +void vluxseg7ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u16mf4))) +void vluxseg7ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u16mf4_m))) +void vluxseg7ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i16m1))) +void vluxseg8ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i16m1_m))) +void vluxseg8ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i16mf2))) +void vluxseg8ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i16mf2_m))) +void vluxseg8ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i16mf4))) +void vluxseg8ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i16mf4_m))) +void vluxseg8ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u16m1))) +void vluxseg8ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u16m1_m))) +void vluxseg8ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u16mf2))) +void vluxseg8ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u16mf2_m))) +void vluxseg8ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u16mf4))) +void vluxseg8ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u16mf4_m))) +void vluxseg8ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32m1))) +void vluxseg2ei8(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32m1_m))) +void vluxseg2ei8(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32m2))) +void vluxseg2ei8(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32m2_m))) +void vluxseg2ei8(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32m4))) +void vluxseg2ei8(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32m4_m))) +void vluxseg2ei8(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32mf2))) +void vluxseg2ei8(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32mf2_m))) +void vluxseg2ei8(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32m1))) +void vluxseg2ei8(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32m1_m))) +void vluxseg2ei8(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32m2))) +void vluxseg2ei8(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32m2_m))) +void vluxseg2ei8(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32m4))) +void vluxseg2ei8(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32m4_m))) +void vluxseg2ei8(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32mf2))) +void vluxseg2ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32mf2_m))) +void vluxseg2ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i32m1))) +void vluxseg3ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i32m1_m))) +void vluxseg3ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i32m2))) +void vluxseg3ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i32m2_m))) +void vluxseg3ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i32mf2))) +void vluxseg3ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i32mf2_m))) +void vluxseg3ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u32m1))) +void vluxseg3ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u32m1_m))) +void vluxseg3ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u32m2))) +void vluxseg3ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u32m2_m))) +void vluxseg3ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u32mf2))) +void vluxseg3ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u32mf2_m))) +void vluxseg3ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i32m1))) +void vluxseg4ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i32m1_m))) +void vluxseg4ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i32m2))) +void vluxseg4ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i32m2_m))) +void vluxseg4ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i32mf2))) +void vluxseg4ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i32mf2_m))) +void vluxseg4ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8m1))) +void vsuxseg2ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8m1_m))) +void vsuxseg2ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8m2))) +void vsuxseg2ei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8m2_m))) +void vsuxseg2ei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8m4))) +void vsuxseg2ei8(int8_t * op0, vuint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8m4_m))) +void vsuxseg2ei8(vbool2_t op0, int8_t * op1, vuint8m4_t op2, vint8m4_t op3, vint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8mf2))) +void vsuxseg2ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8mf2_m))) +void vsuxseg2ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8mf4))) +void vsuxseg2ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8mf4_m))) +void vsuxseg2ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8mf8))) +void vsuxseg2ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8mf8_m))) +void vsuxseg2ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u32m1))) +void vluxseg4ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u32m1_m))) +void vluxseg4ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u32m2))) +void vluxseg4ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u32m2_m))) +void vluxseg4ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u32mf2))) +void vluxseg4ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u32mf2_m))) +void vluxseg4ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i32m1))) +void vluxseg5ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i32m1_m))) +void vluxseg5ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i32mf2))) +void vluxseg5ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i32mf2_m))) +void vluxseg5ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u32m1))) +void vluxseg5ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u32m1_m))) +void vluxseg5ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u32mf2))) +void vluxseg5ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u32mf2_m))) +void vluxseg5ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i32m1))) +void vluxseg6ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i32m1_m))) +void vluxseg6ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i32mf2))) +void vluxseg6ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i32mf2_m))) +void vluxseg6ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u32m1))) +void vluxseg6ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u32m1_m))) +void vluxseg6ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u32mf2))) +void vluxseg6ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u32mf2_m))) +void vluxseg6ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i32m1))) +void vluxseg7ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i32m1_m))) +void vluxseg7ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i32mf2))) +void vluxseg7ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i32mf2_m))) +void vluxseg7ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u32m1))) +void vluxseg7ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u32m1_m))) +void vluxseg7ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u32mf2))) +void vluxseg7ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u32mf2_m))) +void vluxseg7ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i32m1))) +void vluxseg8ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i32m1_m))) +void vluxseg8ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i32mf2))) +void vluxseg8ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i32mf2_m))) +void vluxseg8ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u32m1))) +void vluxseg8ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u32m1_m))) +void vluxseg8ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u32mf2))) +void vluxseg8ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u32mf2_m))) +void vluxseg8ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32m1))) +void vluxseg2ei16(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32m1_m))) +void vluxseg2ei16(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32m2))) +void vluxseg2ei16(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32m2_m))) +void vluxseg2ei16(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32m4))) +void vluxseg2ei16(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32m4_m))) +void vluxseg2ei16(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32mf2))) +void vluxseg2ei16(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32mf2_m))) +void vluxseg2ei16(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8m1))) +void vsuxseg2ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8m1_m))) +void vsuxseg2ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8m2))) +void vsuxseg2ei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8m2_m))) +void vsuxseg2ei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8m4))) +void vsuxseg2ei8(uint8_t * op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8m4_m))) +void vsuxseg2ei8(vbool2_t op0, uint8_t * op1, vuint8m4_t op2, vuint8m4_t op3, vuint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8mf2))) +void vsuxseg2ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8mf2_m))) +void vsuxseg2ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8mf4))) +void vsuxseg2ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8mf4_m))) +void vsuxseg2ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8mf8))) +void vsuxseg2ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8mf8_m))) +void vsuxseg2ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32m1))) +void vluxseg2ei16(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32m1_m))) +void vluxseg2ei16(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32m2))) +void vluxseg2ei16(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32m2_m))) +void vluxseg2ei16(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32m4))) +void vluxseg2ei16(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32m4_m))) +void vluxseg2ei16(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32mf2))) +void vluxseg2ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32mf2_m))) +void vluxseg2ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i32m1))) +void vluxseg3ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i32m1_m))) +void vluxseg3ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i32m2))) +void vluxseg3ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i32m2_m))) +void vluxseg3ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i32mf2))) +void vluxseg3ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i32mf2_m))) +void vluxseg3ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u32m1))) +void vluxseg3ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u32m1_m))) +void vluxseg3ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u32m2))) +void vluxseg3ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u32m2_m))) +void vluxseg3ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u32mf2))) +void vluxseg3ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u32mf2_m))) +void vluxseg3ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i32m1))) +void vluxseg4ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i32m1_m))) +void vluxseg4ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i32m2))) +void vluxseg4ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i32m2_m))) +void vluxseg4ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i32mf2))) +void vluxseg4ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i32mf2_m))) +void vluxseg4ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u32m1))) +void vluxseg4ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u32m1_m))) +void vluxseg4ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u32m2))) +void vluxseg4ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u32m2_m))) +void vluxseg4ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u32mf2))) +void vluxseg4ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u32mf2_m))) +void vluxseg4ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i32m1))) +void vluxseg5ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i32m1_m))) +void vluxseg5ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i32mf2))) +void vluxseg5ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i32mf2_m))) +void vluxseg5ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u32m1))) +void vluxseg5ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u32m1_m))) +void vluxseg5ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u32mf2))) +void vluxseg5ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u32mf2_m))) +void vluxseg5ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i32m1))) +void vluxseg6ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i32m1_m))) +void vluxseg6ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i32mf2))) +void vluxseg6ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i32mf2_m))) +void vluxseg6ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u32m1))) +void vluxseg6ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u32m1_m))) +void vluxseg6ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u32mf2))) +void vluxseg6ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u32mf2_m))) +void vluxseg6ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i32m1))) +void vluxseg7ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i32m1_m))) +void vluxseg7ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i32mf2))) +void vluxseg7ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i32mf2_m))) +void vluxseg7ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u32m1))) +void vluxseg7ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u32m1_m))) +void vluxseg7ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u32mf2))) +void vluxseg7ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u32mf2_m))) +void vluxseg7ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i32m1))) +void vluxseg8ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i32m1_m))) +void vluxseg8ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i32mf2))) +void vluxseg8ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i32mf2_m))) +void vluxseg8ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u32m1))) +void vluxseg8ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u32m1_m))) +void vluxseg8ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u32mf2))) +void vluxseg8ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u32mf2_m))) +void vluxseg8ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32m1))) +void vluxseg2ei32(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32m1_m))) +void vluxseg2ei32(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32m2))) +void vluxseg2ei32(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32m2_m))) +void vluxseg2ei32(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32m4))) +void vluxseg2ei32(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32m4_m))) +void vluxseg2ei32(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32mf2))) +void vluxseg2ei32(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32mf2_m))) +void vluxseg2ei32(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32m1))) +void vluxseg2ei32(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32m1_m))) +void vluxseg2ei32(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32m2))) +void vluxseg2ei32(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32m2_m))) +void vluxseg2ei32(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32m4))) +void vluxseg2ei32(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32m4_m))) +void vluxseg2ei32(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32mf2))) +void vluxseg2ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32mf2_m))) +void vluxseg2ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i32m1))) +void vluxseg3ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i32m1_m))) +void vluxseg3ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i32m2))) +void vluxseg3ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i32m2_m))) +void vluxseg3ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i32mf2))) +void vluxseg3ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i32mf2_m))) +void vluxseg3ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u32m1))) +void vluxseg3ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u32m1_m))) +void vluxseg3ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u32m2))) +void vluxseg3ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u32m2_m))) +void vluxseg3ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u32mf2))) +void vluxseg3ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u32mf2_m))) +void vluxseg3ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i32m1))) +void vluxseg4ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i32m1_m))) +void vluxseg4ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i32m2))) +void vluxseg4ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i32m2_m))) +void vluxseg4ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i32mf2))) +void vluxseg4ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i32mf2_m))) +void vluxseg4ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u32m1))) +void vluxseg4ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u32m1_m))) +void vluxseg4ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u32m2))) +void vluxseg4ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u32m2_m))) +void vluxseg4ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u32mf2))) +void vluxseg4ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u32mf2_m))) +void vluxseg4ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i32m1))) +void vluxseg5ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i32m1_m))) +void vluxseg5ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i32mf2))) +void vluxseg5ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i32mf2_m))) +void vluxseg5ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u32m1))) +void vluxseg5ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u32m1_m))) +void vluxseg5ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u32mf2))) +void vluxseg5ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u32mf2_m))) +void vluxseg5ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i32m1))) +void vluxseg6ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i32m1_m))) +void vluxseg6ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i32mf2))) +void vluxseg6ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i32mf2_m))) +void vluxseg6ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u32m1))) +void vluxseg6ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u32m1_m))) +void vluxseg6ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u32mf2))) +void vluxseg6ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u32mf2_m))) +void vluxseg6ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i32m1))) +void vluxseg7ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i32m1_m))) +void vluxseg7ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i32mf2))) +void vluxseg7ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i32mf2_m))) +void vluxseg7ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u32m1))) +void vluxseg7ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u32m1_m))) +void vluxseg7ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u32mf2))) +void vluxseg7ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u32mf2_m))) +void vluxseg7ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i32m1))) +void vluxseg8ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i32m1_m))) +void vluxseg8ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i32mf2))) +void vluxseg8ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i32mf2_m))) +void vluxseg8ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u32m1))) +void vluxseg8ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u32m1_m))) +void vluxseg8ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u32mf2))) +void vluxseg8ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u32mf2_m))) +void vluxseg8ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32m1))) +void vluxseg2ei64(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32m1_m))) +void vluxseg2ei64(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32m2))) +void vluxseg2ei64(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32m2_m))) +void vluxseg2ei64(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32m4))) +void vluxseg2ei64(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32m4_m))) +void vluxseg2ei64(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32mf2))) +void vluxseg2ei64(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32mf2_m))) +void vluxseg2ei64(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32m1))) +void vluxseg2ei64(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32m1_m))) +void vluxseg2ei64(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32m2))) +void vluxseg2ei64(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32m2_m))) +void vluxseg2ei64(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32m4))) +void vluxseg2ei64(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32m4_m))) +void vluxseg2ei64(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32mf2))) +void vluxseg2ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32mf2_m))) +void vluxseg2ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i32m1))) +void vluxseg3ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i32m1_m))) +void vluxseg3ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i32m2))) +void vluxseg3ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i32m2_m))) +void vluxseg3ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i32mf2))) +void vluxseg3ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i32mf2_m))) +void vluxseg3ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u32m1))) +void vluxseg3ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u32m1_m))) +void vluxseg3ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u32m2))) +void vluxseg3ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u32m2_m))) +void vluxseg3ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u32mf2))) +void vluxseg3ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u32mf2_m))) +void vluxseg3ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i32m1))) +void vluxseg4ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i32m1_m))) +void vluxseg4ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i32m2))) +void vluxseg4ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i32m2_m))) +void vluxseg4ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i32mf2))) +void vluxseg4ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i32mf2_m))) +void vluxseg4ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u32m1))) +void vluxseg4ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u32m1_m))) +void vluxseg4ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u32m2))) +void vluxseg4ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u32m2_m))) +void vluxseg4ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u32mf2))) +void vluxseg4ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u32mf2_m))) +void vluxseg4ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i32m1))) +void vluxseg5ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i32m1_m))) +void vluxseg5ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i32mf2))) +void vluxseg5ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i32mf2_m))) +void vluxseg5ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u32m1))) +void vluxseg5ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u32m1_m))) +void vluxseg5ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u32mf2))) +void vluxseg5ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u32mf2_m))) +void vluxseg5ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i32m1))) +void vluxseg6ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i32m1_m))) +void vluxseg6ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i32mf2))) +void vluxseg6ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i32mf2_m))) +void vluxseg6ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u32m1))) +void vluxseg6ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u32m1_m))) +void vluxseg6ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u32mf2))) +void vluxseg6ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u32mf2_m))) +void vluxseg6ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i32m1))) +void vluxseg7ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i32m1_m))) +void vluxseg7ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i32mf2))) +void vluxseg7ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i32mf2_m))) +void vluxseg7ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u32m1))) +void vluxseg7ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u32m1_m))) +void vluxseg7ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u32mf2))) +void vluxseg7ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u32mf2_m))) +void vluxseg7ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i32m1))) +void vluxseg8ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i32m1_m))) +void vluxseg8ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i32mf2))) +void vluxseg8ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i32mf2_m))) +void vluxseg8ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u32m1))) +void vluxseg8ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u32m1_m))) +void vluxseg8ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u32mf2))) +void vluxseg8ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u32mf2_m))) +void vluxseg8ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i64m1))) +void vluxseg2ei8(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i64m1_m))) +void vluxseg2ei8(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i64m2))) +void vluxseg2ei8(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i64m2_m))) +void vluxseg2ei8(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i64m4))) +void vluxseg2ei8(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i64m4_m))) +void vluxseg2ei8(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u64m1))) +void vluxseg2ei8(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u64m1_m))) +void vluxseg2ei8(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u64m2))) +void vluxseg2ei8(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u64m2_m))) +void vluxseg2ei8(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u64m4))) +void vluxseg2ei8(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u64m4_m))) +void vluxseg2ei8(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i64m1))) +void vluxseg3ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i64m1_m))) +void vluxseg3ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i64m2))) +void vluxseg3ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i64m2_m))) +void vluxseg3ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u64m1))) +void vluxseg3ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u64m1_m))) +void vluxseg3ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u64m2))) +void vluxseg3ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u64m2_m))) +void vluxseg3ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i64m1))) +void vluxseg4ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i64m1_m))) +void vluxseg4ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i64m2))) +void vluxseg4ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i64m2_m))) +void vluxseg4ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u64m1))) +void vluxseg4ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u64m1_m))) +void vluxseg4ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u64m2))) +void vluxseg4ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u64m2_m))) +void vluxseg4ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i64m1))) +void vluxseg5ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i64m1_m))) +void vluxseg5ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u64m1))) +void vluxseg5ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u64m1_m))) +void vluxseg5ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i64m1))) +void vluxseg6ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i64m1_m))) +void vluxseg6ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u64m1))) +void vluxseg6ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u64m1_m))) +void vluxseg6ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i64m1))) +void vluxseg7ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i64m1_m))) +void vluxseg7ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u64m1))) +void vluxseg7ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u64m1_m))) +void vluxseg7ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i64m1))) +void vluxseg8ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i64m1_m))) +void vluxseg8ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u64m1))) +void vluxseg8ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u64m1_m))) +void vluxseg8ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i64m1))) +void vluxseg2ei16(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i64m1_m))) +void vluxseg2ei16(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i64m2))) +void vluxseg2ei16(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i64m2_m))) +void vluxseg2ei16(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i64m4))) +void vluxseg2ei16(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i64m4_m))) +void vluxseg2ei16(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u64m1))) +void vluxseg2ei16(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u64m1_m))) +void vluxseg2ei16(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u64m2))) +void vluxseg2ei16(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u64m2_m))) +void vluxseg2ei16(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u64m4))) +void vluxseg2ei16(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u64m4_m))) +void vluxseg2ei16(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i64m1))) +void vluxseg3ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i64m1_m))) +void vluxseg3ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i64m2))) +void vluxseg3ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i64m2_m))) +void vluxseg3ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u64m1))) +void vluxseg3ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u64m1_m))) +void vluxseg3ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u64m2))) +void vluxseg3ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u64m2_m))) +void vluxseg3ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i64m1))) +void vluxseg4ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i64m1_m))) +void vluxseg4ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i64m2))) +void vluxseg4ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i64m2_m))) +void vluxseg4ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u64m1))) +void vluxseg4ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u64m1_m))) +void vluxseg4ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u64m2))) +void vluxseg4ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u64m2_m))) +void vluxseg4ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i64m1))) +void vluxseg5ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i64m1_m))) +void vluxseg5ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u64m1))) +void vluxseg5ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u64m1_m))) +void vluxseg5ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i64m1))) +void vluxseg6ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i64m1_m))) +void vluxseg6ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u64m1))) +void vluxseg6ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u64m1_m))) +void vluxseg6ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i64m1))) +void vluxseg7ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i64m1_m))) +void vluxseg7ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u64m1))) +void vluxseg7ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u64m1_m))) +void vluxseg7ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i64m1))) +void vluxseg8ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i64m1_m))) +void vluxseg8ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u64m1))) +void vluxseg8ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u64m1_m))) +void vluxseg8ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i64m1))) +void vluxseg2ei32(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i64m1_m))) +void vluxseg2ei32(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i64m2))) +void vluxseg2ei32(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i64m2_m))) +void vluxseg2ei32(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i64m4))) +void vluxseg2ei32(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i64m4_m))) +void vluxseg2ei32(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u64m1))) +void vluxseg2ei32(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u64m1_m))) +void vluxseg2ei32(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u64m2))) +void vluxseg2ei32(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u64m2_m))) +void vluxseg2ei32(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u64m4))) +void vluxseg2ei32(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u64m4_m))) +void vluxseg2ei32(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i64m1))) +void vluxseg3ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i64m1_m))) +void vluxseg3ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i64m2))) +void vluxseg3ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i64m2_m))) +void vluxseg3ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u64m1))) +void vluxseg3ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u64m1_m))) +void vluxseg3ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u64m2))) +void vluxseg3ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u64m2_m))) +void vluxseg3ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i64m1))) +void vluxseg4ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i64m1_m))) +void vluxseg4ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i64m2))) +void vluxseg4ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i64m2_m))) +void vluxseg4ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u64m1))) +void vluxseg4ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u64m1_m))) +void vluxseg4ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u64m2))) +void vluxseg4ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u64m2_m))) +void vluxseg4ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i64m1))) +void vluxseg5ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i64m1_m))) +void vluxseg5ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u64m1))) +void vluxseg5ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u64m1_m))) +void vluxseg5ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i64m1))) +void vluxseg6ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i64m1_m))) +void vluxseg6ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u64m1))) +void vluxseg6ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u64m1_m))) +void vluxseg6ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i64m1))) +void vluxseg7ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i64m1_m))) +void vluxseg7ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u64m1))) +void vluxseg7ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u64m1_m))) +void vluxseg7ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i64m1))) +void vluxseg8ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i64m1_m))) +void vluxseg8ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u64m1))) +void vluxseg8ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u64m1_m))) +void vluxseg8ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i64m1))) +void vluxseg2ei64(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i64m1_m))) +void vluxseg2ei64(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i64m2))) +void vluxseg2ei64(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i64m2_m))) +void vluxseg2ei64(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i64m4))) +void vluxseg2ei64(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i64m4_m))) +void vluxseg2ei64(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u64m1))) +void vluxseg2ei64(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u64m1_m))) +void vluxseg2ei64(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u64m2))) +void vluxseg2ei64(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u64m2_m))) +void vluxseg2ei64(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u64m4))) +void vluxseg2ei64(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u64m4_m))) +void vluxseg2ei64(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i64m1))) +void vluxseg3ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i64m1_m))) +void vluxseg3ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i64m2))) +void vluxseg3ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i64m2_m))) +void vluxseg3ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u64m1))) +void vluxseg3ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u64m1_m))) +void vluxseg3ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u64m2))) +void vluxseg3ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u64m2_m))) +void vluxseg3ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i64m1))) +void vluxseg4ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i64m1_m))) +void vluxseg4ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i64m2))) +void vluxseg4ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i64m2_m))) +void vluxseg4ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u64m1))) +void vluxseg4ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u64m1_m))) +void vluxseg4ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u64m2))) +void vluxseg4ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u64m2_m))) +void vluxseg4ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i64m1))) +void vluxseg5ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i64m1_m))) +void vluxseg5ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u64m1))) +void vluxseg5ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u64m1_m))) +void vluxseg5ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i64m1))) +void vluxseg6ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i64m1_m))) +void vluxseg6ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u64m1))) +void vluxseg6ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u64m1_m))) +void vluxseg6ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i64m1))) +void vluxseg7ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i64m1_m))) +void vluxseg7ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u64m1))) +void vluxseg7ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u64m1_m))) +void vluxseg7ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i64m1))) +void vluxseg8ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i64m1_m))) +void vluxseg8ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u64m1))) +void vluxseg8ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u64m1_m))) +void vluxseg8ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8m1))) +void vloxseg2ei8(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8m1_m))) +void vloxseg2ei8(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8m2))) +void vloxseg2ei8(vint8m2_t * op0, vint8m2_t * op1, const int8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8m2_m))) +void vloxseg2ei8(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8m4))) +void vloxseg2ei8(vint8m4_t * op0, vint8m4_t * op1, const int8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8m4_m))) +void vloxseg2ei8(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, vuint8m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8mf2))) +void vloxseg2ei8(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8mf2_m))) +void vloxseg2ei8(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8mf4))) +void vloxseg2ei8(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8mf4_m))) +void vloxseg2ei8(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8mf8))) +void vloxseg2ei8(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8mf8_m))) +void vloxseg2ei8(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8m1))) +void vloxseg2ei8(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8m1_m))) +void vloxseg2ei8(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8m2))) +void vloxseg2ei8(vuint8m2_t * op0, vuint8m2_t * op1, const uint8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8m2_m))) +void vloxseg2ei8(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8m4))) +void vloxseg2ei8(vuint8m4_t * op0, vuint8m4_t * op1, const uint8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8m4_m))) +void vloxseg2ei8(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, vuint8m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8mf2))) +void vloxseg2ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8mf2_m))) +void vloxseg2ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8mf4))) +void vloxseg2ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8mf4_m))) +void vloxseg2ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8mf8))) +void vloxseg2ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8mf8_m))) +void vloxseg2ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8m1))) +void vloxseg3ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8m1_m))) +void vloxseg3ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8m2))) +void vloxseg3ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, const int8_t * op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8m2_m))) +void vloxseg3ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, vuint8m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8mf2))) +void vloxseg3ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8mf2_m))) +void vloxseg3ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8mf4))) +void vloxseg3ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8mf4_m))) +void vloxseg3ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8mf8))) +void vloxseg3ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8mf8_m))) +void vloxseg3ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8m1))) +void vloxseg3ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8m1_m))) +void vloxseg3ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8m2))) +void vloxseg3ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, const uint8_t * op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8m2_m))) +void vloxseg3ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, vuint8m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8mf2))) +void vloxseg3ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8mf2_m))) +void vloxseg3ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8mf4))) +void vloxseg3ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8mf4_m))) +void vloxseg3ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8mf8))) +void vloxseg3ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8mf8_m))) +void vloxseg3ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8m1))) +void vloxseg4ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8m1_m))) +void vloxseg4ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8m2))) +void vloxseg4ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, const int8_t * op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8m2_m))) +void vloxseg4ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, vuint8m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8mf2))) +void vloxseg4ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8mf2_m))) +void vloxseg4ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8mf4))) +void vloxseg4ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8mf4_m))) +void vloxseg4ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8mf8))) +void vloxseg4ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8mf8_m))) +void vloxseg4ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8m1))) +void vloxseg4ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8m1_m))) +void vloxseg4ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8m2))) +void vloxseg4ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, const uint8_t * op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8m2_m))) +void vloxseg4ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, vuint8m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8mf2))) +void vloxseg4ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8mf2_m))) +void vloxseg4ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8mf4))) +void vloxseg4ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8mf4_m))) +void vloxseg4ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8mf8))) +void vloxseg4ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8mf8_m))) +void vloxseg4ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8m1))) +void vloxseg5ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8m1_m))) +void vloxseg5ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint8m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8mf2))) +void vloxseg5ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8mf2_m))) +void vloxseg5ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8mf4))) +void vloxseg5ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8mf4_m))) +void vloxseg5ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8mf8))) +void vloxseg5ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8mf8_m))) +void vloxseg5ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8m1))) +void vloxseg5ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8m1_m))) +void vloxseg5ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint8m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8mf2))) +void vloxseg5ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8mf2_m))) +void vloxseg5ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8mf4))) +void vloxseg5ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8mf4_m))) +void vloxseg5ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8mf8))) +void vloxseg5ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8mf8_m))) +void vloxseg5ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8m1))) +void vloxseg6ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8m1_m))) +void vloxseg6ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint8m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8mf2))) +void vloxseg6ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8mf2_m))) +void vloxseg6ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8mf4))) +void vloxseg6ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8mf4_m))) +void vloxseg6ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8mf8))) +void vloxseg6ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8mf8_m))) +void vloxseg6ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8m1))) +void vloxseg6ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8m1_m))) +void vloxseg6ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint8m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8mf2))) +void vloxseg6ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8mf2_m))) +void vloxseg6ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8mf4))) +void vloxseg6ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8mf4_m))) +void vloxseg6ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8mf8))) +void vloxseg6ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8mf8_m))) +void vloxseg6ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8m1))) +void vloxseg7ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8m1_m))) +void vloxseg7ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint8m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8mf2))) +void vloxseg7ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8mf2_m))) +void vloxseg7ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8mf4))) +void vloxseg7ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8mf4_m))) +void vloxseg7ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8mf8))) +void vloxseg7ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8mf8_m))) +void vloxseg7ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8m1))) +void vloxseg7ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8m1_m))) +void vloxseg7ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint8m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8mf2))) +void vloxseg7ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8mf2_m))) +void vloxseg7ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8mf4))) +void vloxseg7ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8mf4_m))) +void vloxseg7ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8mf8))) +void vloxseg7ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8mf8_m))) +void vloxseg7ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8m1))) +void vloxseg8ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8m1_m))) +void vloxseg8ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint8m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8mf2))) +void vloxseg8ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8mf2_m))) +void vloxseg8ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8mf4))) +void vloxseg8ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8mf4_m))) +void vloxseg8ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8mf8))) +void vloxseg8ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8mf8_m))) +void vloxseg8ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8m1))) +void vloxseg8ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8m1_m))) +void vloxseg8ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint8m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8mf2))) +void vloxseg8ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8mf2_m))) +void vloxseg8ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8mf4))) +void vloxseg8ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8mf4_m))) +void vloxseg8ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8mf8))) +void vloxseg8ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8mf8_m))) +void vloxseg8ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8m1))) +void vloxseg2ei16(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8m1_m))) +void vloxseg2ei16(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8m2))) +void vloxseg2ei16(vint8m2_t * op0, vint8m2_t * op1, const int8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8m2_m))) +void vloxseg2ei16(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8m4))) +void vloxseg2ei16(vint8m4_t * op0, vint8m4_t * op1, const int8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8m4_m))) +void vloxseg2ei16(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, vuint16m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8mf2))) +void vloxseg2ei16(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8mf2_m))) +void vloxseg2ei16(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8mf4))) +void vloxseg2ei16(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8mf4_m))) +void vloxseg2ei16(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8mf8))) +void vloxseg2ei16(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8mf8_m))) +void vloxseg2ei16(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8m1))) +void vloxseg2ei16(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8m1_m))) +void vloxseg2ei16(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8m2))) +void vloxseg2ei16(vuint8m2_t * op0, vuint8m2_t * op1, const uint8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8m2_m))) +void vloxseg2ei16(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8m4))) +void vloxseg2ei16(vuint8m4_t * op0, vuint8m4_t * op1, const uint8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8m4_m))) +void vloxseg2ei16(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, vuint16m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8mf2))) +void vloxseg2ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8mf2_m))) +void vloxseg2ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8mf4))) +void vloxseg2ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8mf4_m))) +void vloxseg2ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8mf8))) +void vloxseg2ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8mf8_m))) +void vloxseg2ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8m1))) +void vloxseg3ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8m1_m))) +void vloxseg3ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8m2))) +void vloxseg3ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, const int8_t * op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8m2_m))) +void vloxseg3ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, vuint16m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8mf2))) +void vloxseg3ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8mf2_m))) +void vloxseg3ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8mf4))) +void vloxseg3ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8mf4_m))) +void vloxseg3ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8mf8))) +void vloxseg3ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8mf8_m))) +void vloxseg3ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8m1))) +void vloxseg3ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8m1_m))) +void vloxseg3ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8m2))) +void vloxseg3ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, const uint8_t * op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8m2_m))) +void vloxseg3ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, vuint16m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8mf2))) +void vloxseg3ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8mf2_m))) +void vloxseg3ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8mf4))) +void vloxseg3ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8mf4_m))) +void vloxseg3ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8mf8))) +void vloxseg3ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8mf8_m))) +void vloxseg3ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8m1))) +void vloxseg4ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8m1_m))) +void vloxseg4ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8m2))) +void vloxseg4ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, const int8_t * op4, vuint16m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8m2_m))) +void vloxseg4ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, vuint16m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8mf2))) +void vloxseg4ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8mf2_m))) +void vloxseg4ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8mf4))) +void vloxseg4ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8mf4_m))) +void vloxseg4ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8mf8))) +void vloxseg4ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8mf8_m))) +void vloxseg4ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8m1))) +void vloxseg4ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8m1_m))) +void vloxseg4ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8m2))) +void vloxseg4ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, const uint8_t * op4, vuint16m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8m2_m))) +void vloxseg4ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, vuint16m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8mf2))) +void vloxseg4ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8mf2_m))) +void vloxseg4ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8mf4))) +void vloxseg4ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8mf4_m))) +void vloxseg4ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8mf8))) +void vloxseg4ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8mf8_m))) +void vloxseg4ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8m1))) +void vloxseg5ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8m1_m))) +void vloxseg5ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint16m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8mf2))) +void vloxseg5ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8mf2_m))) +void vloxseg5ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8mf4))) +void vloxseg5ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8mf4_m))) +void vloxseg5ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8mf8))) +void vloxseg5ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8mf8_m))) +void vloxseg5ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8m1))) +void vloxseg5ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8m1_m))) +void vloxseg5ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint16m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8mf2))) +void vloxseg5ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8mf2_m))) +void vloxseg5ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8mf4))) +void vloxseg5ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8mf4_m))) +void vloxseg5ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8mf8))) +void vloxseg5ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8mf8_m))) +void vloxseg5ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8m1))) +void vloxseg6ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint16m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8m1_m))) +void vloxseg6ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint16m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8mf2))) +void vloxseg6ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8mf2_m))) +void vloxseg6ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8mf4))) +void vloxseg6ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8mf4_m))) +void vloxseg6ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8mf8))) +void vloxseg6ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8mf8_m))) +void vloxseg6ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8m1))) +void vloxseg6ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint16m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8m1_m))) +void vloxseg6ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint16m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8mf2))) +void vloxseg6ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8mf2_m))) +void vloxseg6ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8mf4))) +void vloxseg6ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8mf4_m))) +void vloxseg6ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8mf8))) +void vloxseg6ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8mf8_m))) +void vloxseg6ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8m1))) +void vloxseg7ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8m1_m))) +void vloxseg7ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint16m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8mf2))) +void vloxseg7ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8mf2_m))) +void vloxseg7ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8mf4))) +void vloxseg7ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8mf4_m))) +void vloxseg7ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8mf8))) +void vloxseg7ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8mf8_m))) +void vloxseg7ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8m1))) +void vloxseg7ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8m1_m))) +void vloxseg7ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint16m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8mf2))) +void vloxseg7ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8mf2_m))) +void vloxseg7ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8mf4))) +void vloxseg7ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8mf4_m))) +void vloxseg7ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8mf8))) +void vloxseg7ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8mf8_m))) +void vloxseg7ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8m1))) +void vloxseg8ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint16m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8m1_m))) +void vloxseg8ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint16m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8mf2))) +void vloxseg8ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8mf2_m))) +void vloxseg8ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8mf4))) +void vloxseg8ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8mf4_m))) +void vloxseg8ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8mf8))) +void vloxseg8ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8mf8_m))) +void vloxseg8ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8m1))) +void vloxseg8ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint16m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8m1_m))) +void vloxseg8ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint16m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8mf2))) +void vloxseg8ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8mf2_m))) +void vloxseg8ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8mf4))) +void vloxseg8ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8mf4_m))) +void vloxseg8ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8mf8))) +void vloxseg8ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8mf8_m))) +void vloxseg8ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8m1))) +void vloxseg2ei32(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8m1_m))) +void vloxseg2ei32(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8m2))) +void vloxseg2ei32(vint8m2_t * op0, vint8m2_t * op1, const int8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8m2_m))) +void vloxseg2ei32(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8mf2))) +void vloxseg2ei32(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8mf2_m))) +void vloxseg2ei32(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8mf4))) +void vloxseg2ei32(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8mf4_m))) +void vloxseg2ei32(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8mf8))) +void vloxseg2ei32(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8mf8_m))) +void vloxseg2ei32(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8m1))) +void vloxseg2ei32(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8m1_m))) +void vloxseg2ei32(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8m2))) +void vloxseg2ei32(vuint8m2_t * op0, vuint8m2_t * op1, const uint8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8m2_m))) +void vloxseg2ei32(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8mf2))) +void vloxseg2ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8mf2_m))) +void vloxseg2ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8mf4))) +void vloxseg2ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8mf4_m))) +void vloxseg2ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8mf8))) +void vloxseg2ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8mf8_m))) +void vloxseg2ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8m1))) +void vloxseg3ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8m1_m))) +void vloxseg3ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8m2))) +void vloxseg3ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, const int8_t * op3, vuint32m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8m2_m))) +void vloxseg3ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, vuint32m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8mf2))) +void vloxseg3ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8mf2_m))) +void vloxseg3ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8mf4))) +void vloxseg3ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8mf4_m))) +void vloxseg3ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8mf8))) +void vloxseg3ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8mf8_m))) +void vloxseg3ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8m1))) +void vloxseg3ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8m1_m))) +void vloxseg3ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8m2))) +void vloxseg3ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, const uint8_t * op3, vuint32m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8m2_m))) +void vloxseg3ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, vuint32m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8mf2))) +void vloxseg3ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8mf2_m))) +void vloxseg3ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8mf4))) +void vloxseg3ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8mf4_m))) +void vloxseg3ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8mf8))) +void vloxseg3ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8mf8_m))) +void vloxseg3ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8m1))) +void vloxseg4ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8m1_m))) +void vloxseg4ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8m2))) +void vloxseg4ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, const int8_t * op4, vuint32m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8m2_m))) +void vloxseg4ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, vuint32m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8mf2))) +void vloxseg4ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8mf2_m))) +void vloxseg4ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8mf4))) +void vloxseg4ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8mf4_m))) +void vloxseg4ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8mf8))) +void vloxseg4ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8mf8_m))) +void vloxseg4ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8m1))) +void vloxseg4ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8m1_m))) +void vloxseg4ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8m2))) +void vloxseg4ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, const uint8_t * op4, vuint32m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8m2_m))) +void vloxseg4ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, vuint32m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8mf2))) +void vloxseg4ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8mf2_m))) +void vloxseg4ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8mf4))) +void vloxseg4ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8mf4_m))) +void vloxseg4ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8mf8))) +void vloxseg4ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8mf8_m))) +void vloxseg4ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8m1))) +void vloxseg5ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8m1_m))) +void vloxseg5ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint32m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8mf2))) +void vloxseg5ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8mf2_m))) +void vloxseg5ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8mf4))) +void vloxseg5ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8mf4_m))) +void vloxseg5ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8mf8))) +void vloxseg5ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8mf8_m))) +void vloxseg5ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8m1))) +void vloxseg5ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8m1_m))) +void vloxseg5ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint32m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8mf2))) +void vloxseg5ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8mf2_m))) +void vloxseg5ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8mf4))) +void vloxseg5ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8mf4_m))) +void vloxseg5ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8mf8))) +void vloxseg5ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8mf8_m))) +void vloxseg5ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8m1))) +void vloxseg6ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint32m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8m1_m))) +void vloxseg6ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint32m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8mf2))) +void vloxseg6ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8mf2_m))) +void vloxseg6ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8mf4))) +void vloxseg6ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8mf4_m))) +void vloxseg6ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8mf8))) +void vloxseg6ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8mf8_m))) +void vloxseg6ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8m1))) +void vloxseg6ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint32m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8m1_m))) +void vloxseg6ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint32m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8mf2))) +void vloxseg6ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8mf2_m))) +void vloxseg6ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8mf4))) +void vloxseg6ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8mf4_m))) +void vloxseg6ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8mf8))) +void vloxseg6ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8mf8_m))) +void vloxseg6ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8m1))) +void vloxseg7ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8m1_m))) +void vloxseg7ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint32m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8mf2))) +void vloxseg7ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8mf2_m))) +void vloxseg7ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8mf4))) +void vloxseg7ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8mf4_m))) +void vloxseg7ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8mf8))) +void vloxseg7ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8mf8_m))) +void vloxseg7ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8m1))) +void vloxseg7ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8m1_m))) +void vloxseg7ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint32m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8mf2))) +void vloxseg7ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8mf2_m))) +void vloxseg7ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8mf4))) +void vloxseg7ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8mf4_m))) +void vloxseg7ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8mf8))) +void vloxseg7ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8mf8_m))) +void vloxseg7ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8m1))) +void vloxseg8ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint32m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8m1_m))) +void vloxseg8ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint32m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8mf2))) +void vloxseg8ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8mf2_m))) +void vloxseg8ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8mf4))) +void vloxseg8ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8mf4_m))) +void vloxseg8ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8mf8))) +void vloxseg8ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8mf8_m))) +void vloxseg8ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8m1))) +void vloxseg8ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint32m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8m1_m))) +void vloxseg8ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint32m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8mf2))) +void vloxseg8ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8mf2_m))) +void vloxseg8ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8mf4))) +void vloxseg8ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8mf4_m))) +void vloxseg8ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8mf8))) +void vloxseg8ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8mf8_m))) +void vloxseg8ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8m1))) +void vloxseg2ei64(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8m1_m))) +void vloxseg2ei64(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8mf2))) +void vloxseg2ei64(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8mf2_m))) +void vloxseg2ei64(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8mf4))) +void vloxseg2ei64(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8mf4_m))) +void vloxseg2ei64(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8mf8))) +void vloxseg2ei64(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8mf8_m))) +void vloxseg2ei64(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8m1))) +void vloxseg2ei64(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8m1_m))) +void vloxseg2ei64(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8mf2))) +void vloxseg2ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8mf2_m))) +void vloxseg2ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8mf4))) +void vloxseg2ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8mf4_m))) +void vloxseg2ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8mf8))) +void vloxseg2ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8mf8_m))) +void vloxseg2ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8m1))) +void vloxseg3ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8m1_m))) +void vloxseg3ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8mf2))) +void vloxseg3ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8mf2_m))) +void vloxseg3ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8mf4))) +void vloxseg3ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8mf4_m))) +void vloxseg3ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8mf8))) +void vloxseg3ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8mf8_m))) +void vloxseg3ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8m1))) +void vloxseg3ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8m1_m))) +void vloxseg3ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8mf2))) +void vloxseg3ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8mf2_m))) +void vloxseg3ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8mf4))) +void vloxseg3ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8mf4_m))) +void vloxseg3ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8mf8))) +void vloxseg3ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8mf8_m))) +void vloxseg3ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8m1))) +void vloxseg4ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8m1_m))) +void vloxseg4ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8mf2))) +void vloxseg4ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8mf2_m))) +void vloxseg4ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8mf4))) +void vloxseg4ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8mf4_m))) +void vloxseg4ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8mf8))) +void vloxseg4ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8mf8_m))) +void vloxseg4ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8m1))) +void vloxseg4ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8m1_m))) +void vloxseg4ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8mf2))) +void vloxseg4ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8mf2_m))) +void vloxseg4ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8mf4))) +void vloxseg4ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8mf4_m))) +void vloxseg4ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8mf8))) +void vloxseg4ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8mf8_m))) +void vloxseg4ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8m1))) +void vloxseg5ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8m1_m))) +void vloxseg5ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint64m8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8mf2))) +void vloxseg5ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8mf2_m))) +void vloxseg5ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8mf4))) +void vloxseg5ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8mf4_m))) +void vloxseg5ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8mf8))) +void vloxseg5ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8mf8_m))) +void vloxseg5ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8m1))) +void vloxseg5ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8m1_m))) +void vloxseg5ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint64m8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8mf2))) +void vloxseg5ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8mf2_m))) +void vloxseg5ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8mf4))) +void vloxseg5ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8mf4_m))) +void vloxseg5ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8mf8))) +void vloxseg5ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8mf8_m))) +void vloxseg5ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8m1))) +void vloxseg6ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint64m8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8m1_m))) +void vloxseg6ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint64m8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8mf2))) +void vloxseg6ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8mf2_m))) +void vloxseg6ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8mf4))) +void vloxseg6ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8mf4_m))) +void vloxseg6ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8mf8))) +void vloxseg6ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8mf8_m))) +void vloxseg6ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8m1))) +void vloxseg6ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint64m8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8m1_m))) +void vloxseg6ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint64m8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8mf2))) +void vloxseg6ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8mf2_m))) +void vloxseg6ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8mf4))) +void vloxseg6ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8mf4_m))) +void vloxseg6ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8mf8))) +void vloxseg6ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8mf8_m))) +void vloxseg6ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8m1))) +void vloxseg7ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8m1_m))) +void vloxseg7ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint64m8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8mf2))) +void vloxseg7ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8mf2_m))) +void vloxseg7ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8mf4))) +void vloxseg7ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8mf4_m))) +void vloxseg7ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8mf8))) +void vloxseg7ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8mf8_m))) +void vloxseg7ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8m1))) +void vloxseg7ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8m1_m))) +void vloxseg7ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint64m8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8mf2))) +void vloxseg7ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8mf2_m))) +void vloxseg7ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8mf4))) +void vloxseg7ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8mf4_m))) +void vloxseg7ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8mf8))) +void vloxseg7ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8mf8_m))) +void vloxseg7ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8m1))) +void vloxseg8ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint64m8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8m1_m))) +void vloxseg8ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint64m8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8mf2))) +void vloxseg8ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8mf2_m))) +void vloxseg8ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8mf4))) +void vloxseg8ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8mf4_m))) +void vloxseg8ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8mf8))) +void vloxseg8ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8mf8_m))) +void vloxseg8ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8m1))) +void vloxseg8ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint64m8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8m1_m))) +void vloxseg8ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint64m8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8mf2))) +void vloxseg8ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8mf2_m))) +void vloxseg8ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8mf4))) +void vloxseg8ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8mf4_m))) +void vloxseg8ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8mf8))) +void vloxseg8ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8mf8_m))) +void vloxseg8ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16m1))) +void vloxseg2ei8(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16m1_m))) +void vloxseg2ei8(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16m2))) +void vloxseg2ei8(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16m2_m))) +void vloxseg2ei8(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16m4))) +void vloxseg2ei8(vint16m4_t * op0, vint16m4_t * op1, const int16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16m4_m))) +void vloxseg2ei8(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16mf2))) +void vloxseg2ei8(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16mf2_m))) +void vloxseg2ei8(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16mf4))) +void vloxseg2ei8(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16mf4_m))) +void vloxseg2ei8(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16m1))) +void vloxseg2ei8(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16m1_m))) +void vloxseg2ei8(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16m2))) +void vloxseg2ei8(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16m2_m))) +void vloxseg2ei8(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16m4))) +void vloxseg2ei8(vuint16m4_t * op0, vuint16m4_t * op1, const uint16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16m4_m))) +void vloxseg2ei8(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16mf2))) +void vloxseg2ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16mf2_m))) +void vloxseg2ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16mf4))) +void vloxseg2ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16mf4_m))) +void vloxseg2ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16m1))) +void vloxseg3ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16m1_m))) +void vloxseg3ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16m2))) +void vloxseg3ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16m2_m))) +void vloxseg3ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16mf2))) +void vloxseg3ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16mf2_m))) +void vloxseg3ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16mf4))) +void vloxseg3ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16mf4_m))) +void vloxseg3ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16m1))) +void vloxseg3ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16m1_m))) +void vloxseg3ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16m2))) +void vloxseg3ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16m2_m))) +void vloxseg3ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16mf2))) +void vloxseg3ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16mf2_m))) +void vloxseg3ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16mf4))) +void vloxseg3ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16mf4_m))) +void vloxseg3ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16m1))) +void vloxseg4ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16m1_m))) +void vloxseg4ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16m2))) +void vloxseg4ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16m2_m))) +void vloxseg4ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16mf2))) +void vloxseg4ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16mf2_m))) +void vloxseg4ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16mf4))) +void vloxseg4ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16mf4_m))) +void vloxseg4ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16m1))) +void vloxseg4ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16m1_m))) +void vloxseg4ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16m2))) +void vloxseg4ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16m2_m))) +void vloxseg4ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16mf2))) +void vloxseg4ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16mf2_m))) +void vloxseg4ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16mf4))) +void vloxseg4ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16mf4_m))) +void vloxseg4ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i16m1))) +void vloxseg5ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i16m1_m))) +void vloxseg5ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i16mf2))) +void vloxseg5ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i16mf2_m))) +void vloxseg5ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i16mf4))) +void vloxseg5ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i16mf4_m))) +void vloxseg5ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u16m1))) +void vloxseg5ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u16m1_m))) +void vloxseg5ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u16mf2))) +void vloxseg5ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u16mf2_m))) +void vloxseg5ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u16mf4))) +void vloxseg5ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u16mf4_m))) +void vloxseg5ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i16m1))) +void vloxseg6ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i16m1_m))) +void vloxseg6ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i16mf2))) +void vloxseg6ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i16mf2_m))) +void vloxseg6ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i16mf4))) +void vloxseg6ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i16mf4_m))) +void vloxseg6ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u16m1))) +void vloxseg6ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u16m1_m))) +void vloxseg6ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u16mf2))) +void vloxseg6ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u16mf2_m))) +void vloxseg6ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u16mf4))) +void vloxseg6ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u16mf4_m))) +void vloxseg6ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i16m1))) +void vloxseg7ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i16m1_m))) +void vloxseg7ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i16mf2))) +void vloxseg7ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i16mf2_m))) +void vloxseg7ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i16mf4))) +void vloxseg7ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i16mf4_m))) +void vloxseg7ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u16m1))) +void vloxseg7ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u16m1_m))) +void vloxseg7ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u16mf2))) +void vloxseg7ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u16mf2_m))) +void vloxseg7ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u16mf4))) +void vloxseg7ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u16mf4_m))) +void vloxseg7ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i16m1))) +void vloxseg8ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i16m1_m))) +void vloxseg8ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i16mf2))) +void vloxseg8ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i16mf2_m))) +void vloxseg8ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i16mf4))) +void vloxseg8ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i16mf4_m))) +void vloxseg8ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u16m1))) +void vloxseg8ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u16m1_m))) +void vloxseg8ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u16mf2))) +void vloxseg8ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u16mf2_m))) +void vloxseg8ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u16mf4))) +void vloxseg8ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u16mf4_m))) +void vloxseg8ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16m1))) +void vloxseg2ei16(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16m1_m))) +void vloxseg2ei16(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16m2))) +void vloxseg2ei16(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16m2_m))) +void vloxseg2ei16(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16m4))) +void vloxseg2ei16(vint16m4_t * op0, vint16m4_t * op1, const int16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16m4_m))) +void vloxseg2ei16(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16mf2))) +void vloxseg2ei16(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16mf2_m))) +void vloxseg2ei16(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16mf4))) +void vloxseg2ei16(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16mf4_m))) +void vloxseg2ei16(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16m1))) +void vloxseg2ei16(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16m1_m))) +void vloxseg2ei16(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16m2))) +void vloxseg2ei16(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16m2_m))) +void vloxseg2ei16(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16m4))) +void vloxseg2ei16(vuint16m4_t * op0, vuint16m4_t * op1, const uint16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16m4_m))) +void vloxseg2ei16(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16mf2))) +void vloxseg2ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16mf2_m))) +void vloxseg2ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16mf4))) +void vloxseg2ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16mf4_m))) +void vloxseg2ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16m1))) +void vloxseg3ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16m1_m))) +void vloxseg3ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16m2))) +void vloxseg3ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16m2_m))) +void vloxseg3ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16mf2))) +void vloxseg3ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16mf2_m))) +void vloxseg3ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16mf4))) +void vloxseg3ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16mf4_m))) +void vloxseg3ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16m1))) +void vloxseg3ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16m1_m))) +void vloxseg3ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16m2))) +void vloxseg3ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16m2_m))) +void vloxseg3ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16mf2))) +void vloxseg3ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16mf2_m))) +void vloxseg3ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16mf4))) +void vloxseg3ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16mf4_m))) +void vloxseg3ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16m1))) +void vloxseg4ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16m1_m))) +void vloxseg4ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16m2))) +void vloxseg4ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16m2_m))) +void vloxseg4ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16mf2))) +void vloxseg4ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16mf2_m))) +void vloxseg4ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16mf4))) +void vloxseg4ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16mf4_m))) +void vloxseg4ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16m1))) +void vloxseg4ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16m1_m))) +void vloxseg4ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16m2))) +void vloxseg4ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16m2_m))) +void vloxseg4ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16mf2))) +void vloxseg4ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16mf2_m))) +void vloxseg4ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16mf4))) +void vloxseg4ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16mf4_m))) +void vloxseg4ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i16m1))) +void vloxseg5ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i16m1_m))) +void vloxseg5ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i16mf2))) +void vloxseg5ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i16mf2_m))) +void vloxseg5ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i16mf4))) +void vloxseg5ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i16mf4_m))) +void vloxseg5ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u16m1))) +void vloxseg5ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u16m1_m))) +void vloxseg5ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u16mf2))) +void vloxseg5ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u16mf2_m))) +void vloxseg5ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u16mf4))) +void vloxseg5ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u16mf4_m))) +void vloxseg5ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i16m1))) +void vloxseg6ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i16m1_m))) +void vloxseg6ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i16mf2))) +void vloxseg6ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i16mf2_m))) +void vloxseg6ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i16mf4))) +void vloxseg6ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i16mf4_m))) +void vloxseg6ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u16m1))) +void vloxseg6ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u16m1_m))) +void vloxseg6ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u16mf2))) +void vloxseg6ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u16mf2_m))) +void vloxseg6ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u16mf4))) +void vloxseg6ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u16mf4_m))) +void vloxseg6ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i16m1))) +void vloxseg7ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i16m1_m))) +void vloxseg7ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i16mf2))) +void vloxseg7ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i16mf2_m))) +void vloxseg7ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i16mf4))) +void vloxseg7ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i16mf4_m))) +void vloxseg7ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u16m1))) +void vloxseg7ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u16m1_m))) +void vloxseg7ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u16mf2))) +void vloxseg7ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u16mf2_m))) +void vloxseg7ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u16mf4))) +void vloxseg7ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u16mf4_m))) +void vloxseg7ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i16m1))) +void vloxseg8ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i16m1_m))) +void vloxseg8ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i16mf2))) +void vloxseg8ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i16mf2_m))) +void vloxseg8ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i16mf4))) +void vloxseg8ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i16mf4_m))) +void vloxseg8ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u16m1))) +void vloxseg8ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u16m1_m))) +void vloxseg8ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u16mf2))) +void vloxseg8ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u16mf2_m))) +void vloxseg8ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u16mf4))) +void vloxseg8ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u16mf4_m))) +void vloxseg8ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16m1))) +void vloxseg2ei32(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16m1_m))) +void vloxseg2ei32(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16m2))) +void vloxseg2ei32(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16m2_m))) +void vloxseg2ei32(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16m4))) +void vloxseg2ei32(vint16m4_t * op0, vint16m4_t * op1, const int16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16m4_m))) +void vloxseg2ei32(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16mf2))) +void vloxseg2ei32(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16mf2_m))) +void vloxseg2ei32(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16mf4))) +void vloxseg2ei32(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16mf4_m))) +void vloxseg2ei32(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16m1))) +void vloxseg2ei32(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16m1_m))) +void vloxseg2ei32(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16m2))) +void vloxseg2ei32(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16m2_m))) +void vloxseg2ei32(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16m4))) +void vloxseg2ei32(vuint16m4_t * op0, vuint16m4_t * op1, const uint16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16m4_m))) +void vloxseg2ei32(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16mf2))) +void vloxseg2ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16mf2_m))) +void vloxseg2ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16mf4))) +void vloxseg2ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16mf4_m))) +void vloxseg2ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16m1))) +void vloxseg3ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16m1_m))) +void vloxseg3ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16m2))) +void vloxseg3ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16m2_m))) +void vloxseg3ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16mf2))) +void vloxseg3ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16mf2_m))) +void vloxseg3ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16mf4))) +void vloxseg3ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16mf4_m))) +void vloxseg3ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16m1))) +void vloxseg3ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16m1_m))) +void vloxseg3ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16m2))) +void vloxseg3ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16m2_m))) +void vloxseg3ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16mf2))) +void vloxseg3ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16mf2_m))) +void vloxseg3ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16mf4))) +void vloxseg3ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16mf4_m))) +void vloxseg3ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16m1))) +void vloxseg4ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16m1_m))) +void vloxseg4ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16m2))) +void vloxseg4ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16m2_m))) +void vloxseg4ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16mf2))) +void vloxseg4ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16mf2_m))) +void vloxseg4ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16mf4))) +void vloxseg4ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16mf4_m))) +void vloxseg4ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16m1))) +void vloxseg4ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16m1_m))) +void vloxseg4ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16m2))) +void vloxseg4ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16m2_m))) +void vloxseg4ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16mf2))) +void vloxseg4ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16mf2_m))) +void vloxseg4ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16mf4))) +void vloxseg4ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16mf4_m))) +void vloxseg4ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i16m1))) +void vloxseg5ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i16m1_m))) +void vloxseg5ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i16mf2))) +void vloxseg5ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i16mf2_m))) +void vloxseg5ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i16mf4))) +void vloxseg5ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i16mf4_m))) +void vloxseg5ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u16m1))) +void vloxseg5ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u16m1_m))) +void vloxseg5ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u16mf2))) +void vloxseg5ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u16mf2_m))) +void vloxseg5ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u16mf4))) +void vloxseg5ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u16mf4_m))) +void vloxseg5ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i16m1))) +void vloxseg6ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i16m1_m))) +void vloxseg6ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i16mf2))) +void vloxseg6ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i16mf2_m))) +void vloxseg6ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i16mf4))) +void vloxseg6ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i16mf4_m))) +void vloxseg6ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u16m1))) +void vloxseg6ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u16m1_m))) +void vloxseg6ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u16mf2))) +void vloxseg6ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u16mf2_m))) +void vloxseg6ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u16mf4))) +void vloxseg6ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u16mf4_m))) +void vloxseg6ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i16m1))) +void vloxseg7ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i16m1_m))) +void vloxseg7ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i16mf2))) +void vloxseg7ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i16mf2_m))) +void vloxseg7ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i16mf4))) +void vloxseg7ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i16mf4_m))) +void vloxseg7ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u16m1))) +void vloxseg7ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u16m1_m))) +void vloxseg7ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u16mf2))) +void vloxseg7ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u16mf2_m))) +void vloxseg7ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u16mf4))) +void vloxseg7ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u16mf4_m))) +void vloxseg7ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i16m1))) +void vloxseg8ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i16m1_m))) +void vloxseg8ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i16mf2))) +void vloxseg8ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i16mf2_m))) +void vloxseg8ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i16mf4))) +void vloxseg8ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i16mf4_m))) +void vloxseg8ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u16m1))) +void vloxseg8ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u16m1_m))) +void vloxseg8ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u16mf2))) +void vloxseg8ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u16mf2_m))) +void vloxseg8ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u16mf4))) +void vloxseg8ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u16mf4_m))) +void vloxseg8ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16m1))) +void vloxseg2ei64(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16m1_m))) +void vloxseg2ei64(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16m2))) +void vloxseg2ei64(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16m2_m))) +void vloxseg2ei64(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16mf2))) +void vloxseg2ei64(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16mf2_m))) +void vloxseg2ei64(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16mf4))) +void vloxseg2ei64(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16mf4_m))) +void vloxseg2ei64(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16m1))) +void vloxseg2ei64(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16m1_m))) +void vloxseg2ei64(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16m2))) +void vloxseg2ei64(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16m2_m))) +void vloxseg2ei64(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16mf2))) +void vloxseg2ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16mf2_m))) +void vloxseg2ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16mf4))) +void vloxseg2ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16mf4_m))) +void vloxseg2ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16m1))) +void vloxseg3ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16m1_m))) +void vloxseg3ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16m2))) +void vloxseg3ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16m2_m))) +void vloxseg3ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16mf2))) +void vloxseg3ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16mf2_m))) +void vloxseg3ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16mf4))) +void vloxseg3ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16mf4_m))) +void vloxseg3ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16m1))) +void vloxseg3ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16m1_m))) +void vloxseg3ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16m2))) +void vloxseg3ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16m2_m))) +void vloxseg3ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16mf2))) +void vloxseg3ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16mf2_m))) +void vloxseg3ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16mf4))) +void vloxseg3ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16mf4_m))) +void vloxseg3ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16m1))) +void vloxseg4ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16m1_m))) +void vloxseg4ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16m2))) +void vloxseg4ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16m2_m))) +void vloxseg4ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16mf2))) +void vloxseg4ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16mf2_m))) +void vloxseg4ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16mf4))) +void vloxseg4ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16mf4_m))) +void vloxseg4ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16m1))) +void vloxseg4ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16m1_m))) +void vloxseg4ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16m2))) +void vloxseg4ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16m2_m))) +void vloxseg4ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16mf2))) +void vloxseg4ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16mf2_m))) +void vloxseg4ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16mf4))) +void vloxseg4ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16mf4_m))) +void vloxseg4ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i16m1))) +void vloxseg5ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i16m1_m))) +void vloxseg5ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i16mf2))) +void vloxseg5ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i16mf2_m))) +void vloxseg5ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i16mf4))) +void vloxseg5ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i16mf4_m))) +void vloxseg5ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u16m1))) +void vloxseg5ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u16m1_m))) +void vloxseg5ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u16mf2))) +void vloxseg5ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u16mf2_m))) +void vloxseg5ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u16mf4))) +void vloxseg5ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u16mf4_m))) +void vloxseg5ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i16m1))) +void vloxseg6ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i16m1_m))) +void vloxseg6ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i16mf2))) +void vloxseg6ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i16mf2_m))) +void vloxseg6ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i16mf4))) +void vloxseg6ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i16mf4_m))) +void vloxseg6ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u16m1))) +void vloxseg6ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u16m1_m))) +void vloxseg6ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u16mf2))) +void vloxseg6ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u16mf2_m))) +void vloxseg6ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u16mf4))) +void vloxseg6ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u16mf4_m))) +void vloxseg6ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i16m1))) +void vloxseg7ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i16m1_m))) +void vloxseg7ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i16mf2))) +void vloxseg7ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i16mf2_m))) +void vloxseg7ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i16mf4))) +void vloxseg7ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i16mf4_m))) +void vloxseg7ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u16m1))) +void vloxseg7ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u16m1_m))) +void vloxseg7ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u16mf2))) +void vloxseg7ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u16mf2_m))) +void vloxseg7ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u16mf4))) +void vloxseg7ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u16mf4_m))) +void vloxseg7ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i16m1))) +void vloxseg8ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i16m1_m))) +void vloxseg8ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i16mf2))) +void vloxseg8ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i16mf2_m))) +void vloxseg8ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i16mf4))) +void vloxseg8ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i16mf4_m))) +void vloxseg8ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u16m1))) +void vloxseg8ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u16m1_m))) +void vloxseg8ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u16mf2))) +void vloxseg8ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u16mf2_m))) +void vloxseg8ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u16mf4))) +void vloxseg8ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u16mf4_m))) +void vloxseg8ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32m1))) +void vloxseg2ei8(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32m1_m))) +void vloxseg2ei8(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32m2))) +void vloxseg2ei8(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32m2_m))) +void vloxseg2ei8(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32m4))) +void vloxseg2ei8(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32m4_m))) +void vloxseg2ei8(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32mf2))) +void vloxseg2ei8(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32mf2_m))) +void vloxseg2ei8(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32m1))) +void vloxseg2ei8(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32m1_m))) +void vloxseg2ei8(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32m2))) +void vloxseg2ei8(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32m2_m))) +void vloxseg2ei8(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32m4))) +void vloxseg2ei8(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32m4_m))) +void vloxseg2ei8(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32mf2))) +void vloxseg2ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32mf2_m))) +void vloxseg2ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i32m1))) +void vloxseg3ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i32m1_m))) +void vloxseg3ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i32m2))) +void vloxseg3ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i32m2_m))) +void vloxseg3ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i32mf2))) +void vloxseg3ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i32mf2_m))) +void vloxseg3ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u32m1))) +void vloxseg3ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u32m1_m))) +void vloxseg3ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u32m2))) +void vloxseg3ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u32m2_m))) +void vloxseg3ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u32mf2))) +void vloxseg3ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u32mf2_m))) +void vloxseg3ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i32m1))) +void vloxseg4ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i32m1_m))) +void vloxseg4ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i32m2))) +void vloxseg4ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i32m2_m))) +void vloxseg4ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i32mf2))) +void vloxseg4ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i32mf2_m))) +void vloxseg4ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u32m1))) +void vloxseg4ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u32m1_m))) +void vloxseg4ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u32m2))) +void vloxseg4ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u32m2_m))) +void vloxseg4ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u32mf2))) +void vloxseg4ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u32mf2_m))) +void vloxseg4ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i32m1))) +void vloxseg5ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i32m1_m))) +void vloxseg5ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i32mf2))) +void vloxseg5ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i32mf2_m))) +void vloxseg5ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u32m1))) +void vloxseg5ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u32m1_m))) +void vloxseg5ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u32mf2))) +void vloxseg5ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u32mf2_m))) +void vloxseg5ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i32m1))) +void vloxseg6ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i32m1_m))) +void vloxseg6ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i32mf2))) +void vloxseg6ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i32mf2_m))) +void vloxseg6ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u32m1))) +void vloxseg6ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u32m1_m))) +void vloxseg6ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u32mf2))) +void vloxseg6ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u32mf2_m))) +void vloxseg6ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i32m1))) +void vloxseg7ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i32m1_m))) +void vloxseg7ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i32mf2))) +void vloxseg7ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i32mf2_m))) +void vloxseg7ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u32m1))) +void vloxseg7ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u32m1_m))) +void vloxseg7ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u32mf2))) +void vloxseg7ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u32mf2_m))) +void vloxseg7ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i32m1))) +void vloxseg8ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i32m1_m))) +void vloxseg8ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i32mf2))) +void vloxseg8ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i32mf2_m))) +void vloxseg8ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u32m1))) +void vloxseg8ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u32m1_m))) +void vloxseg8ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u32mf2))) +void vloxseg8ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u32mf2_m))) +void vloxseg8ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32m1))) +void vloxseg2ei16(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32m1_m))) +void vloxseg2ei16(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32m2))) +void vloxseg2ei16(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32m2_m))) +void vloxseg2ei16(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32m4))) +void vloxseg2ei16(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32m4_m))) +void vloxseg2ei16(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32mf2))) +void vloxseg2ei16(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32mf2_m))) +void vloxseg2ei16(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32m1))) +void vloxseg2ei16(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32m1_m))) +void vloxseg2ei16(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32m2))) +void vloxseg2ei16(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32m2_m))) +void vloxseg2ei16(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32m4))) +void vloxseg2ei16(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32m4_m))) +void vloxseg2ei16(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32mf2))) +void vloxseg2ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32mf2_m))) +void vloxseg2ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i32m1))) +void vloxseg3ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i32m1_m))) +void vloxseg3ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i32m2))) +void vloxseg3ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i32m2_m))) +void vloxseg3ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i32mf2))) +void vloxseg3ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i32mf2_m))) +void vloxseg3ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u32m1))) +void vloxseg3ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u32m1_m))) +void vloxseg3ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u32m2))) +void vloxseg3ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u32m2_m))) +void vloxseg3ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u32mf2))) +void vloxseg3ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u32mf2_m))) +void vloxseg3ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i32m1))) +void vloxseg4ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i32m1_m))) +void vloxseg4ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i32m2))) +void vloxseg4ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i32m2_m))) +void vloxseg4ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i32mf2))) +void vloxseg4ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i32mf2_m))) +void vloxseg4ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u32m1))) +void vloxseg4ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u32m1_m))) +void vloxseg4ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u32m2))) +void vloxseg4ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u32m2_m))) +void vloxseg4ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u32mf2))) +void vloxseg4ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u32mf2_m))) +void vloxseg4ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i32m1))) +void vloxseg5ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i32m1_m))) +void vloxseg5ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i32mf2))) +void vloxseg5ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i32mf2_m))) +void vloxseg5ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u32m1))) +void vloxseg5ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u32m1_m))) +void vloxseg5ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u32mf2))) +void vloxseg5ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u32mf2_m))) +void vloxseg5ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i32m1))) +void vloxseg6ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i32m1_m))) +void vloxseg6ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i32mf2))) +void vloxseg6ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i32mf2_m))) +void vloxseg6ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u32m1))) +void vloxseg6ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u32m1_m))) +void vloxseg6ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u32mf2))) +void vloxseg6ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u32mf2_m))) +void vloxseg6ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i32m1))) +void vloxseg7ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i32m1_m))) +void vloxseg7ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i32mf2))) +void vloxseg7ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i32mf2_m))) +void vloxseg7ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u32m1))) +void vloxseg7ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u32m1_m))) +void vloxseg7ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u32mf2))) +void vloxseg7ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u32mf2_m))) +void vloxseg7ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i32m1))) +void vloxseg8ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i32m1_m))) +void vloxseg8ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i32mf2))) +void vloxseg8ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i32mf2_m))) +void vloxseg8ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u32m1))) +void vloxseg8ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u32m1_m))) +void vloxseg8ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u32mf2))) +void vloxseg8ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u32mf2_m))) +void vloxseg8ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32m1))) +void vloxseg2ei32(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32m1_m))) +void vloxseg2ei32(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32m2))) +void vloxseg2ei32(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32m2_m))) +void vloxseg2ei32(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32m4))) +void vloxseg2ei32(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32m4_m))) +void vloxseg2ei32(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32mf2))) +void vloxseg2ei32(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32mf2_m))) +void vloxseg2ei32(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32m1))) +void vloxseg2ei32(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32m1_m))) +void vloxseg2ei32(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32m2))) +void vloxseg2ei32(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32m2_m))) +void vloxseg2ei32(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32m4))) +void vloxseg2ei32(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32m4_m))) +void vloxseg2ei32(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32mf2))) +void vloxseg2ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32mf2_m))) +void vloxseg2ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i32m1))) +void vloxseg3ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i32m1_m))) +void vloxseg3ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i32m2))) +void vloxseg3ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i32m2_m))) +void vloxseg3ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i32mf2))) +void vloxseg3ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i32mf2_m))) +void vloxseg3ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u32m1))) +void vloxseg3ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u32m1_m))) +void vloxseg3ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u32m2))) +void vloxseg3ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u32m2_m))) +void vloxseg3ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u32mf2))) +void vloxseg3ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u32mf2_m))) +void vloxseg3ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i32m1))) +void vloxseg4ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i32m1_m))) +void vloxseg4ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i32m2))) +void vloxseg4ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i32m2_m))) +void vloxseg4ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i32mf2))) +void vloxseg4ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i32mf2_m))) +void vloxseg4ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u32m1))) +void vloxseg4ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u32m1_m))) +void vloxseg4ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u32m2))) +void vloxseg4ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u32m2_m))) +void vloxseg4ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u32mf2))) +void vloxseg4ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u32mf2_m))) +void vloxseg4ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i32m1))) +void vloxseg5ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i32m1_m))) +void vloxseg5ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i32mf2))) +void vloxseg5ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i32mf2_m))) +void vloxseg5ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u32m1))) +void vloxseg5ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u32m1_m))) +void vloxseg5ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u32mf2))) +void vloxseg5ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u32mf2_m))) +void vloxseg5ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i32m1))) +void vloxseg6ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i32m1_m))) +void vloxseg6ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i32mf2))) +void vloxseg6ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i32mf2_m))) +void vloxseg6ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u32m1))) +void vloxseg6ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u32m1_m))) +void vloxseg6ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u32mf2))) +void vloxseg6ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u32mf2_m))) +void vloxseg6ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i32m1))) +void vloxseg7ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i32m1_m))) +void vloxseg7ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i32mf2))) +void vloxseg7ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i32mf2_m))) +void vloxseg7ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u32m1))) +void vloxseg7ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u32m1_m))) +void vloxseg7ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u32mf2))) +void vloxseg7ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u32mf2_m))) +void vloxseg7ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i32m1))) +void vloxseg8ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i32m1_m))) +void vloxseg8ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i32mf2))) +void vloxseg8ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i32mf2_m))) +void vloxseg8ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u32m1))) +void vloxseg8ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u32m1_m))) +void vloxseg8ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u32mf2))) +void vloxseg8ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u32mf2_m))) +void vloxseg8ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32m1))) +void vloxseg2ei64(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32m1_m))) +void vloxseg2ei64(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32m2))) +void vloxseg2ei64(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32m2_m))) +void vloxseg2ei64(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32m4))) +void vloxseg2ei64(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32m4_m))) +void vloxseg2ei64(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32mf2))) +void vloxseg2ei64(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32mf2_m))) +void vloxseg2ei64(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32m1))) +void vloxseg2ei64(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32m1_m))) +void vloxseg2ei64(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32m2))) +void vloxseg2ei64(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32m2_m))) +void vloxseg2ei64(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32m4))) +void vloxseg2ei64(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32m4_m))) +void vloxseg2ei64(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32mf2))) +void vloxseg2ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32mf2_m))) +void vloxseg2ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i32m1))) +void vloxseg3ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i32m1_m))) +void vloxseg3ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i32m2))) +void vloxseg3ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i32m2_m))) +void vloxseg3ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i32mf2))) +void vloxseg3ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i32mf2_m))) +void vloxseg3ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u32m1))) +void vloxseg3ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u32m1_m))) +void vloxseg3ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u32m2))) +void vloxseg3ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u32m2_m))) +void vloxseg3ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u32mf2))) +void vloxseg3ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u32mf2_m))) +void vloxseg3ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i32m1))) +void vloxseg4ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i32m1_m))) +void vloxseg4ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i32m2))) +void vloxseg4ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i32m2_m))) +void vloxseg4ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i32mf2))) +void vloxseg4ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i32mf2_m))) +void vloxseg4ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u32m1))) +void vloxseg4ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u32m1_m))) +void vloxseg4ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u32m2))) +void vloxseg4ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u32m2_m))) +void vloxseg4ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u32mf2))) +void vloxseg4ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u32mf2_m))) +void vloxseg4ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i32m1))) +void vloxseg5ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i32m1_m))) +void vloxseg5ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i32mf2))) +void vloxseg5ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i32mf2_m))) +void vloxseg5ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u32m1))) +void vloxseg5ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u32m1_m))) +void vloxseg5ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u32mf2))) +void vloxseg5ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u32mf2_m))) +void vloxseg5ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i32m1))) +void vloxseg6ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i32m1_m))) +void vloxseg6ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i32mf2))) +void vloxseg6ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i32mf2_m))) +void vloxseg6ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u32m1))) +void vloxseg6ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u32m1_m))) +void vloxseg6ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u32mf2))) +void vloxseg6ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u32mf2_m))) +void vloxseg6ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i32m1))) +void vloxseg7ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i32m1_m))) +void vloxseg7ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i32mf2))) +void vloxseg7ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i32mf2_m))) +void vloxseg7ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u32m1))) +void vloxseg7ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u32m1_m))) +void vloxseg7ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u32mf2))) +void vloxseg7ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u32mf2_m))) +void vloxseg7ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i32m1))) +void vloxseg8ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i32m1_m))) +void vloxseg8ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i32mf2))) +void vloxseg8ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i32mf2_m))) +void vloxseg8ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u32m1))) +void vloxseg8ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u32m1_m))) +void vloxseg8ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u32mf2))) +void vloxseg8ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u32mf2_m))) +void vloxseg8ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint64m1_t op18, size_t op19); + +#endif + +#if defined(__riscv_f) && defined(__riscv_zvlsseg) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32m1))) +void vloxseg2ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32m1_m))) +void vloxseg2ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32m2))) +void vloxseg2ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32m2_m))) +void vloxseg2ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32m4))) +void vloxseg2ei8(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32m4_m))) +void vloxseg2ei8(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32mf2))) +void vloxseg2ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32mf2_m))) +void vloxseg2ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f32m1))) +void vloxseg3ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f32m1_m))) +void vloxseg3ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f32m2))) +void vloxseg3ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f32m2_m))) +void vloxseg3ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f32mf2))) +void vloxseg3ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f32mf2_m))) +void vloxseg3ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f32m1))) +void vloxseg4ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f32m1_m))) +void vloxseg4ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f32m2))) +void vloxseg4ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f32m2_m))) +void vloxseg4ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f32mf2))) +void vloxseg4ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f32mf2_m))) +void vloxseg4ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f32m1))) +void vloxseg5ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f32m1_m))) +void vloxseg5ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f32mf2))) +void vloxseg5ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f32mf2_m))) +void vloxseg5ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f32m1))) +void vloxseg6ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f32m1_m))) +void vloxseg6ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f32mf2))) +void vloxseg6ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f32mf2_m))) +void vloxseg6ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f32m1))) +void vloxseg7ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f32m1_m))) +void vloxseg7ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f32mf2))) +void vloxseg7ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f32mf2_m))) +void vloxseg7ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f32m1))) +void vloxseg8ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f32m1_m))) +void vloxseg8ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f32mf2))) +void vloxseg8ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f32mf2_m))) +void vloxseg8ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32m1))) +void vloxseg2ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32m1_m))) +void vloxseg2ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32m2))) +void vloxseg2ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32m2_m))) +void vloxseg2ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32m4))) +void vloxseg2ei16(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32m4_m))) +void vloxseg2ei16(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32mf2))) +void vloxseg2ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32mf2_m))) +void vloxseg2ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f32m1))) +void vloxseg3ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f32m1_m))) +void vloxseg3ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f32m2))) +void vloxseg3ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f32m2_m))) +void vloxseg3ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f32mf2))) +void vloxseg3ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f32mf2_m))) +void vloxseg3ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f32m1))) +void vloxseg4ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f32m1_m))) +void vloxseg4ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f32m2))) +void vloxseg4ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f32m2_m))) +void vloxseg4ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f32mf2))) +void vloxseg4ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f32mf2_m))) +void vloxseg4ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f32m1))) +void vloxseg5ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f32m1_m))) +void vloxseg5ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f32mf2))) +void vloxseg5ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f32mf2_m))) +void vloxseg5ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f32m1))) +void vloxseg6ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f32m1_m))) +void vloxseg6ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f32mf2))) +void vloxseg6ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f32mf2_m))) +void vloxseg6ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f32m1))) +void vloxseg7ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f32m1_m))) +void vloxseg7ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f32mf2))) +void vloxseg7ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f32mf2_m))) +void vloxseg7ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f32m1))) +void vloxseg8ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f32m1_m))) +void vloxseg8ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f32mf2))) +void vloxseg8ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f32mf2_m))) +void vloxseg8ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32m1))) +void vloxseg2ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32m1_m))) +void vloxseg2ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32m2))) +void vloxseg2ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32m2_m))) +void vloxseg2ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32m4))) +void vloxseg2ei32(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32m4_m))) +void vloxseg2ei32(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32mf2))) +void vloxseg2ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32mf2_m))) +void vloxseg2ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f32m1))) +void vloxseg3ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f32m1_m))) +void vloxseg3ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f32m2))) +void vloxseg3ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f32m2_m))) +void vloxseg3ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f32mf2))) +void vloxseg3ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f32mf2_m))) +void vloxseg3ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f32m1))) +void vloxseg4ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f32m1_m))) +void vloxseg4ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f32m2))) +void vloxseg4ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f32m2_m))) +void vloxseg4ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f32mf2))) +void vloxseg4ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f32mf2_m))) +void vloxseg4ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f32m1))) +void vloxseg5ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f32m1_m))) +void vloxseg5ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f32mf2))) +void vloxseg5ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f32mf2_m))) +void vloxseg5ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f32m1))) +void vloxseg6ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f32m1_m))) +void vloxseg6ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f32mf2))) +void vloxseg6ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f32mf2_m))) +void vloxseg6ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f32m1))) +void vloxseg7ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f32m1_m))) +void vloxseg7ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f32mf2))) +void vloxseg7ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f32mf2_m))) +void vloxseg7ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f32m1))) +void vloxseg8ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f32m1_m))) +void vloxseg8ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f32mf2))) +void vloxseg8ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f32mf2_m))) +void vloxseg8ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32m1))) +void vloxseg2ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32m1_m))) +void vloxseg2ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32m2))) +void vloxseg2ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32m2_m))) +void vloxseg2ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32m4))) +void vloxseg2ei64(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32m4_m))) +void vloxseg2ei64(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32mf2))) +void vloxseg2ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32mf2_m))) +void vloxseg2ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f32m1))) +void vloxseg3ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f32m1_m))) +void vloxseg3ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f32m2))) +void vloxseg3ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f32m2_m))) +void vloxseg3ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f32mf2))) +void vloxseg3ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f32mf2_m))) +void vloxseg3ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f32m1))) +void vloxseg4ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f32m1_m))) +void vloxseg4ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f32m2))) +void vloxseg4ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f32m2_m))) +void vloxseg4ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f32mf2))) +void vloxseg4ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f32mf2_m))) +void vloxseg4ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f32m1))) +void vloxseg5ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f32m1_m))) +void vloxseg5ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f32mf2))) +void vloxseg5ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f32mf2_m))) +void vloxseg5ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f32m1))) +void vloxseg6ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f32m1_m))) +void vloxseg6ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f32mf2))) +void vloxseg6ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f32mf2_m))) +void vloxseg6ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f32m1))) +void vloxseg7ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f32m1_m))) +void vloxseg7ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f32mf2))) +void vloxseg7ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f32mf2_m))) +void vloxseg7ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f32m1))) +void vloxseg8ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f32m1_m))) +void vloxseg8ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f32mf2))) +void vloxseg8ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f32mf2_m))) +void vloxseg8ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32m1))) +void vsseg2e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32m1_m))) +void vsseg2e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32m2))) +void vsseg2e32(float * op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32m2_m))) +void vsseg2e32(vbool16_t op0, float * op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32m4))) +void vsseg2e32(float * op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32m4_m))) +void vsseg2e32(vbool8_t op0, float * op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32mf2))) +void vsseg2e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32mf2_m))) +void vsseg2e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_f32m1))) +void vsseg3e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_f32m1_m))) +void vsseg3e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_f32m2))) +void vsseg3e32(float * op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_f32m2_m))) +void vsseg3e32(vbool16_t op0, float * op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_f32mf2))) +void vsseg3e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_f32mf2_m))) +void vsseg3e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_f32m1))) +void vsseg4e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_f32m1_m))) +void vsseg4e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_f32m2))) +void vsseg4e32(float * op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_f32m2_m))) +void vsseg4e32(vbool16_t op0, float * op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_f32mf2))) +void vsseg4e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_f32mf2_m))) +void vsseg4e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_f32m1))) +void vsseg5e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_f32m1_m))) +void vsseg5e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_f32mf2))) +void vsseg5e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_f32mf2_m))) +void vsseg5e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_f32m1))) +void vsseg6e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_f32m1_m))) +void vsseg6e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_f32mf2))) +void vsseg6e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_f32mf2_m))) +void vsseg6e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_f32m1))) +void vsseg7e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_f32m1_m))) +void vsseg7e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_f32mf2))) +void vsseg7e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_f32mf2_m))) +void vsseg7e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_f32m1))) +void vsseg8e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_f32m1_m))) +void vsseg8e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_f32mf2))) +void vsseg8e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_f32mf2_m))) +void vsseg8e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32m1))) +void vssseg2e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32m1_m))) +void vssseg2e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32m2))) +void vssseg2e32(float * op0, ptrdiff_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32m2_m))) +void vssseg2e32(vbool16_t op0, float * op1, ptrdiff_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32m4))) +void vssseg2e32(float * op0, ptrdiff_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32m4_m))) +void vssseg2e32(vbool8_t op0, float * op1, ptrdiff_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32mf2))) +void vssseg2e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32mf2_m))) +void vssseg2e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_f32m1))) +void vssseg3e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_f32m1_m))) +void vssseg3e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_f32m2))) +void vssseg3e32(float * op0, ptrdiff_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_f32m2_m))) +void vssseg3e32(vbool16_t op0, float * op1, ptrdiff_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_f32mf2))) +void vssseg3e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_f32mf2_m))) +void vssseg3e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_f32m1))) +void vssseg4e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_f32m1_m))) +void vssseg4e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_f32m2))) +void vssseg4e32(float * op0, ptrdiff_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_f32m2_m))) +void vssseg4e32(vbool16_t op0, float * op1, ptrdiff_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_f32mf2))) +void vssseg4e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_f32mf2_m))) +void vssseg4e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_f32m1))) +void vssseg5e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_f32m1_m))) +void vssseg5e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_f32mf2))) +void vssseg5e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_f32mf2_m))) +void vssseg5e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_f32m1))) +void vssseg6e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_f32m1_m))) +void vssseg6e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_f32mf2))) +void vssseg6e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_f32mf2_m))) +void vssseg6e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_f32m1))) +void vssseg7e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_f32m1_m))) +void vssseg7e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_f32mf2))) +void vssseg7e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_f32mf2_m))) +void vssseg7e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_f32m1))) +void vssseg8e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_f32m1_m))) +void vssseg8e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_f32mf2))) +void vssseg8e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_f32mf2_m))) +void vssseg8e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32m1))) +void vsuxseg2ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32m1_m))) +void vsuxseg2ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32m2))) +void vsuxseg2ei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32m2_m))) +void vsuxseg2ei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32m4))) +void vsuxseg2ei8(float * op0, vuint8m1_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32m4_m))) +void vsuxseg2ei8(vbool8_t op0, float * op1, vuint8m1_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32mf2))) +void vsuxseg2ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32mf2_m))) +void vsuxseg2ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f32m1))) +void vsuxseg3ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f32m1_m))) +void vsuxseg3ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f32m2))) +void vsuxseg3ei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f32m2_m))) +void vsuxseg3ei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f32mf2))) +void vsuxseg3ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f32mf2_m))) +void vsuxseg3ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f32m1))) +void vsuxseg4ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f32m1_m))) +void vsuxseg4ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f32m2))) +void vsuxseg4ei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f32m2_m))) +void vsuxseg4ei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f32mf2))) +void vsuxseg4ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f32mf2_m))) +void vsuxseg4ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f32m1))) +void vsuxseg5ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f32m1_m))) +void vsuxseg5ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f32mf2))) +void vsuxseg5ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f32mf2_m))) +void vsuxseg5ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f32m1))) +void vsuxseg6ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f32m1_m))) +void vsuxseg6ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f32mf2))) +void vsuxseg6ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f32mf2_m))) +void vsuxseg6ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f32m1))) +void vsuxseg7ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f32m1_m))) +void vsuxseg7ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f32mf2))) +void vsuxseg7ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f32mf2_m))) +void vsuxseg7ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f32m1))) +void vsuxseg8ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f32m1_m))) +void vsuxseg8ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f32mf2))) +void vsuxseg8ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f32mf2_m))) +void vsuxseg8ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32m1))) +void vsuxseg2ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32m1_m))) +void vsuxseg2ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32m2))) +void vsuxseg2ei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32m2_m))) +void vsuxseg2ei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32m4))) +void vsuxseg2ei16(float * op0, vuint16m2_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32m4_m))) +void vsuxseg2ei16(vbool8_t op0, float * op1, vuint16m2_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32mf2))) +void vsuxseg2ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32mf2_m))) +void vsuxseg2ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f32m1))) +void vsuxseg3ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f32m1_m))) +void vsuxseg3ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f32m2))) +void vsuxseg3ei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f32m2_m))) +void vsuxseg3ei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f32mf2))) +void vsuxseg3ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f32mf2_m))) +void vsuxseg3ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f32m1))) +void vsuxseg4ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f32m1_m))) +void vsuxseg4ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f32m2))) +void vsuxseg4ei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f32m2_m))) +void vsuxseg4ei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f32mf2))) +void vsuxseg4ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f32mf2_m))) +void vsuxseg4ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f32m1))) +void vsuxseg5ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f32m1_m))) +void vsuxseg5ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f32mf2))) +void vsuxseg5ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f32mf2_m))) +void vsuxseg5ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f32m1))) +void vsuxseg6ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f32m1_m))) +void vsuxseg6ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f32mf2))) +void vsuxseg6ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f32mf2_m))) +void vsuxseg6ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f32m1))) +void vsuxseg7ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f32m1_m))) +void vsuxseg7ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f32mf2))) +void vsuxseg7ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f32mf2_m))) +void vsuxseg7ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f32m1))) +void vsuxseg8ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f32m1_m))) +void vsuxseg8ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f32mf2))) +void vsuxseg8ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f32mf2_m))) +void vsuxseg8ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32m1))) +void vsuxseg2ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32m1_m))) +void vsuxseg2ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32m2))) +void vsuxseg2ei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32m2_m))) +void vsuxseg2ei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32m4))) +void vsuxseg2ei32(float * op0, vuint32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32m4_m))) +void vsuxseg2ei32(vbool8_t op0, float * op1, vuint32m4_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32mf2))) +void vsuxseg2ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32mf2_m))) +void vsuxseg2ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f32m1))) +void vsuxseg3ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f32m1_m))) +void vsuxseg3ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f32m2))) +void vsuxseg3ei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f32m2_m))) +void vsuxseg3ei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f32mf2))) +void vsuxseg3ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f32mf2_m))) +void vsuxseg3ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f32m1))) +void vsuxseg4ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f32m1_m))) +void vsuxseg4ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f32m2))) +void vsuxseg4ei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f32m2_m))) +void vsuxseg4ei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f32mf2))) +void vsuxseg4ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f32mf2_m))) +void vsuxseg4ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f32m1))) +void vsuxseg5ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f32m1_m))) +void vsuxseg5ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f32mf2))) +void vsuxseg5ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f32mf2_m))) +void vsuxseg5ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f32m1))) +void vsuxseg6ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f32m1_m))) +void vsuxseg6ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f32mf2))) +void vsuxseg6ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f32mf2_m))) +void vsuxseg6ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f32m1))) +void vsuxseg7ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f32m1_m))) +void vsuxseg7ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f32mf2))) +void vsuxseg7ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f32mf2_m))) +void vsuxseg7ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f32m1))) +void vsuxseg8ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f32m1_m))) +void vsuxseg8ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f32mf2))) +void vsuxseg8ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f32mf2_m))) +void vsuxseg8ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32m1))) +void vsuxseg2ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32m1_m))) +void vsuxseg2ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32m2))) +void vsuxseg2ei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32m2_m))) +void vsuxseg2ei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32m4))) +void vsuxseg2ei64(float * op0, vuint64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32m4_m))) +void vsuxseg2ei64(vbool8_t op0, float * op1, vuint64m8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32mf2))) +void vsuxseg2ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32mf2_m))) +void vsuxseg2ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f32m1))) +void vsuxseg3ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f32m1_m))) +void vsuxseg3ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f32m2))) +void vsuxseg3ei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f32m2_m))) +void vsuxseg3ei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f32mf2))) +void vsuxseg3ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f32mf2_m))) +void vsuxseg3ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f32m1))) +void vsuxseg4ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f32m1_m))) +void vsuxseg4ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f32m2))) +void vsuxseg4ei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f32m2_m))) +void vsuxseg4ei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f32mf2))) +void vsuxseg4ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f32mf2_m))) +void vsuxseg4ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f32m1))) +void vsuxseg5ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f32m1_m))) +void vsuxseg5ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f32mf2))) +void vsuxseg5ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f32mf2_m))) +void vsuxseg5ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f32m1))) +void vsuxseg6ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f32m1_m))) +void vsuxseg6ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f32mf2))) +void vsuxseg6ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f32mf2_m))) +void vsuxseg6ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f32m1))) +void vsuxseg7ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f32m1_m))) +void vsuxseg7ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f32mf2))) +void vsuxseg7ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f32mf2_m))) +void vsuxseg7ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f32m1))) +void vsuxseg8ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f32m1_m))) +void vsuxseg8ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f32mf2))) +void vsuxseg8ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f32mf2_m))) +void vsuxseg8ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32m1))) +void vsoxseg2ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32m1_m))) +void vsoxseg2ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32m2))) +void vsoxseg2ei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32m2_m))) +void vsoxseg2ei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32m4))) +void vsoxseg2ei8(float * op0, vuint8m1_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32m4_m))) +void vsoxseg2ei8(vbool8_t op0, float * op1, vuint8m1_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32mf2))) +void vsoxseg2ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32mf2_m))) +void vsoxseg2ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f32m1))) +void vsoxseg3ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f32m1_m))) +void vsoxseg3ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f32m2))) +void vsoxseg3ei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f32m2_m))) +void vsoxseg3ei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f32mf2))) +void vsoxseg3ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f32mf2_m))) +void vsoxseg3ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f32m1))) +void vsoxseg4ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f32m1_m))) +void vsoxseg4ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f32m2))) +void vsoxseg4ei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f32m2_m))) +void vsoxseg4ei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f32mf2))) +void vsoxseg4ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f32mf2_m))) +void vsoxseg4ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f32m1))) +void vsoxseg5ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f32m1_m))) +void vsoxseg5ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f32mf2))) +void vsoxseg5ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f32mf2_m))) +void vsoxseg5ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f32m1))) +void vsoxseg6ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f32m1_m))) +void vsoxseg6ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f32mf2))) +void vsoxseg6ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f32mf2_m))) +void vsoxseg6ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f32m1))) +void vsoxseg7ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f32m1_m))) +void vsoxseg7ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f32mf2))) +void vsoxseg7ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f32mf2_m))) +void vsoxseg7ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f32m1))) +void vsoxseg8ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f32m1_m))) +void vsoxseg8ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f32mf2))) +void vsoxseg8ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f32mf2_m))) +void vsoxseg8ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32m1))) +void vsoxseg2ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32m1_m))) +void vsoxseg2ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32m2))) +void vsoxseg2ei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32m2_m))) +void vsoxseg2ei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32m4))) +void vsoxseg2ei16(float * op0, vuint16m2_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32m4_m))) +void vsoxseg2ei16(vbool8_t op0, float * op1, vuint16m2_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32mf2))) +void vsoxseg2ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32mf2_m))) +void vsoxseg2ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f32m1))) +void vsoxseg3ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f32m1_m))) +void vsoxseg3ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f32m2))) +void vsoxseg3ei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f32m2_m))) +void vsoxseg3ei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f32mf2))) +void vsoxseg3ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f32mf2_m))) +void vsoxseg3ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f32m1))) +void vsoxseg4ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f32m1_m))) +void vsoxseg4ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f32m2))) +void vsoxseg4ei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f32m2_m))) +void vsoxseg4ei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f32mf2))) +void vsoxseg4ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f32mf2_m))) +void vsoxseg4ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f32m1))) +void vsoxseg5ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f32m1_m))) +void vsoxseg5ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f32mf2))) +void vsoxseg5ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f32mf2_m))) +void vsoxseg5ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f32m1))) +void vsoxseg6ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f32m1_m))) +void vsoxseg6ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f32mf2))) +void vsoxseg6ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f32mf2_m))) +void vsoxseg6ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f32m1))) +void vsoxseg7ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f32m1_m))) +void vsoxseg7ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f32mf2))) +void vsoxseg7ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f32mf2_m))) +void vsoxseg7ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f32m1))) +void vsoxseg8ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f32m1_m))) +void vsoxseg8ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f32mf2))) +void vsoxseg8ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f32mf2_m))) +void vsoxseg8ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32m1))) +void vsoxseg2ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32m1_m))) +void vsoxseg2ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32m2))) +void vsoxseg2ei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32m2_m))) +void vsoxseg2ei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32m4))) +void vsoxseg2ei32(float * op0, vuint32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32m4_m))) +void vsoxseg2ei32(vbool8_t op0, float * op1, vuint32m4_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32mf2))) +void vsoxseg2ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32mf2_m))) +void vsoxseg2ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f32m1))) +void vsoxseg3ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f32m1_m))) +void vsoxseg3ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f32m2))) +void vsoxseg3ei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f32m2_m))) +void vsoxseg3ei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f32mf2))) +void vsoxseg3ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f32mf2_m))) +void vsoxseg3ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f32m1))) +void vsoxseg4ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f32m1_m))) +void vsoxseg4ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f32m2))) +void vsoxseg4ei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f32m2_m))) +void vsoxseg4ei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f32mf2))) +void vsoxseg4ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f32mf2_m))) +void vsoxseg4ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f32m1))) +void vsoxseg5ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f32m1_m))) +void vsoxseg5ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f32mf2))) +void vsoxseg5ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f32mf2_m))) +void vsoxseg5ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f32m1))) +void vsoxseg6ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f32m1_m))) +void vsoxseg6ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f32mf2))) +void vsoxseg6ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f32mf2_m))) +void vsoxseg6ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f32m1))) +void vsoxseg7ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f32m1_m))) +void vsoxseg7ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f32mf2))) +void vsoxseg7ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f32mf2_m))) +void vsoxseg7ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f32m1))) +void vsoxseg8ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f32m1_m))) +void vsoxseg8ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f32mf2))) +void vsoxseg8ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f32mf2_m))) +void vsoxseg8ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32m1))) +void vsoxseg2ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32m1_m))) +void vsoxseg2ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32m2))) +void vsoxseg2ei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32m2_m))) +void vsoxseg2ei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32m4))) +void vsoxseg2ei64(float * op0, vuint64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32m4_m))) +void vsoxseg2ei64(vbool8_t op0, float * op1, vuint64m8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32mf2))) +void vsoxseg2ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32mf2_m))) +void vsoxseg2ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f32m1))) +void vsoxseg3ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f32m1_m))) +void vsoxseg3ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f32m2))) +void vsoxseg3ei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f32m2_m))) +void vsoxseg3ei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f32mf2))) +void vsoxseg3ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f32mf2_m))) +void vsoxseg3ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f32m1))) +void vsoxseg4ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f32m1_m))) +void vsoxseg4ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f32m2))) +void vsoxseg4ei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f32m2_m))) +void vsoxseg4ei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f32mf2))) +void vsoxseg4ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f32mf2_m))) +void vsoxseg4ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f32m1))) +void vsoxseg5ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f32m1_m))) +void vsoxseg5ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f32mf2))) +void vsoxseg5ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f32mf2_m))) +void vsoxseg5ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f32m1))) +void vsoxseg6ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f32m1_m))) +void vsoxseg6ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f32mf2))) +void vsoxseg6ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f32mf2_m))) +void vsoxseg6ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f32m1))) +void vsoxseg7ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f32m1_m))) +void vsoxseg7ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f32mf2))) +void vsoxseg7ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f32mf2_m))) +void vsoxseg7ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f32m1))) +void vsoxseg8ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f32m1_m))) +void vsoxseg8ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f32mf2))) +void vsoxseg8ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f32mf2_m))) +void vsoxseg8ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_f32m1_m))) +void vlseg2e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_f32m2_m))) +void vlseg2e32(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_f32m4_m))) +void vlseg2e32(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_f32mf2_m))) +void vlseg2e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_f32m1_m))) +void vlseg3e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_f32m2_m))) +void vlseg3e32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_f32mf2_m))) +void vlseg3e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_f32m1_m))) +void vlseg4e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_f32m2_m))) +void vlseg4e32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_f32mf2_m))) +void vlseg4e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32_v_f32m1_m))) +void vlseg5e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32_v_f32mf2_m))) +void vlseg5e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32_v_f32m1_m))) +void vlseg6e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32_v_f32mf2_m))) +void vlseg6e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32_v_f32m1_m))) +void vlseg7e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32_v_f32mf2_m))) +void vlseg7e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32_v_f32m1_m))) +void vlseg8e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32_v_f32mf2_m))) +void vlseg8e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_f32m1_m))) +void vlseg2e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_f32m2_m))) +void vlseg2e32ff(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_f32m4_m))) +void vlseg2e32ff(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_f32mf2_m))) +void vlseg2e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_f32m1_m))) +void vlseg3e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_f32m2_m))) +void vlseg3e32ff(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_f32mf2_m))) +void vlseg3e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_f32m1_m))) +void vlseg4e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_f32m2_m))) +void vlseg4e32ff(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_f32mf2_m))) +void vlseg4e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32ff_v_f32m1_m))) +void vlseg5e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32ff_v_f32mf2_m))) +void vlseg5e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32ff_v_f32m1_m))) +void vlseg6e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32ff_v_f32mf2_m))) +void vlseg6e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32ff_v_f32m1_m))) +void vlseg7e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32ff_v_f32mf2_m))) +void vlseg7e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32ff_v_f32m1_m))) +void vlseg8e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32ff_v_f32mf2_m))) +void vlseg8e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_f32m1_m))) +void vlsseg2e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_f32m2_m))) +void vlsseg2e32(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_f32m4_m))) +void vlsseg2e32(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_f32mf2_m))) +void vlsseg2e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_f32m1_m))) +void vlsseg3e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_f32m2_m))) +void vlsseg3e32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_f32mf2_m))) +void vlsseg3e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_f32m1_m))) +void vlsseg4e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_f32m2_m))) +void vlsseg4e32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_f32mf2_m))) +void vlsseg4e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e32_v_f32m1_m))) +void vlsseg5e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e32_v_f32mf2_m))) +void vlsseg5e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e32_v_f32m1_m))) +void vlsseg6e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e32_v_f32mf2_m))) +void vlsseg6e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e32_v_f32m1_m))) +void vlsseg7e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e32_v_f32mf2_m))) +void vlsseg7e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e32_v_f32m1_m))) +void vlsseg8e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e32_v_f32mf2_m))) +void vlsseg8e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32m1))) +void vluxseg2ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32m1_m))) +void vluxseg2ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32m2))) +void vluxseg2ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32m2_m))) +void vluxseg2ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32m4))) +void vluxseg2ei8(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32m4_m))) +void vluxseg2ei8(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32mf2))) +void vluxseg2ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32mf2_m))) +void vluxseg2ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f32m1))) +void vluxseg3ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f32m1_m))) +void vluxseg3ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f32m2))) +void vluxseg3ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f32m2_m))) +void vluxseg3ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f32mf2))) +void vluxseg3ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f32mf2_m))) +void vluxseg3ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f32m1))) +void vluxseg4ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f32m1_m))) +void vluxseg4ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f32m2))) +void vluxseg4ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f32m2_m))) +void vluxseg4ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f32mf2))) +void vluxseg4ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f32mf2_m))) +void vluxseg4ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f32m1))) +void vluxseg5ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f32m1_m))) +void vluxseg5ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f32mf2))) +void vluxseg5ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f32mf2_m))) +void vluxseg5ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f32m1))) +void vluxseg6ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f32m1_m))) +void vluxseg6ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f32mf2))) +void vluxseg6ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f32mf2_m))) +void vluxseg6ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f32m1))) +void vluxseg7ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f32m1_m))) +void vluxseg7ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f32mf2))) +void vluxseg7ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f32mf2_m))) +void vluxseg7ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f32m1))) +void vluxseg8ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f32m1_m))) +void vluxseg8ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f32mf2))) +void vluxseg8ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f32mf2_m))) +void vluxseg8ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32m1))) +void vluxseg2ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32m1_m))) +void vluxseg2ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32m2))) +void vluxseg2ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32m2_m))) +void vluxseg2ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32m4))) +void vluxseg2ei16(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32m4_m))) +void vluxseg2ei16(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32mf2))) +void vluxseg2ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32mf2_m))) +void vluxseg2ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f32m1))) +void vluxseg3ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f32m1_m))) +void vluxseg3ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f32m2))) +void vluxseg3ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f32m2_m))) +void vluxseg3ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f32mf2))) +void vluxseg3ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f32mf2_m))) +void vluxseg3ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f32m1))) +void vluxseg4ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f32m1_m))) +void vluxseg4ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f32m2))) +void vluxseg4ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f32m2_m))) +void vluxseg4ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f32mf2))) +void vluxseg4ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f32mf2_m))) +void vluxseg4ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f32m1))) +void vluxseg5ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f32m1_m))) +void vluxseg5ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f32mf2))) +void vluxseg5ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f32mf2_m))) +void vluxseg5ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f32m1))) +void vluxseg6ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f32m1_m))) +void vluxseg6ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f32mf2))) +void vluxseg6ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f32mf2_m))) +void vluxseg6ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f32m1))) +void vluxseg7ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f32m1_m))) +void vluxseg7ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f32mf2))) +void vluxseg7ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f32mf2_m))) +void vluxseg7ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f32m1))) +void vluxseg8ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f32m1_m))) +void vluxseg8ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f32mf2))) +void vluxseg8ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f32mf2_m))) +void vluxseg8ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32m1))) +void vluxseg2ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32m1_m))) +void vluxseg2ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32m2))) +void vluxseg2ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32m2_m))) +void vluxseg2ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32m4))) +void vluxseg2ei32(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32m4_m))) +void vluxseg2ei32(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32mf2))) +void vluxseg2ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32mf2_m))) +void vluxseg2ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f32m1))) +void vluxseg3ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f32m1_m))) +void vluxseg3ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f32m2))) +void vluxseg3ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f32m2_m))) +void vluxseg3ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f32mf2))) +void vluxseg3ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f32mf2_m))) +void vluxseg3ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f32m1))) +void vluxseg4ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f32m1_m))) +void vluxseg4ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f32m2))) +void vluxseg4ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f32m2_m))) +void vluxseg4ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f32mf2))) +void vluxseg4ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f32mf2_m))) +void vluxseg4ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f32m1))) +void vluxseg5ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f32m1_m))) +void vluxseg5ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f32mf2))) +void vluxseg5ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f32mf2_m))) +void vluxseg5ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f32m1))) +void vluxseg6ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f32m1_m))) +void vluxseg6ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f32mf2))) +void vluxseg6ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f32mf2_m))) +void vluxseg6ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f32m1))) +void vluxseg7ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f32m1_m))) +void vluxseg7ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f32mf2))) +void vluxseg7ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f32mf2_m))) +void vluxseg7ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f32m1))) +void vluxseg8ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f32m1_m))) +void vluxseg8ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f32mf2))) +void vluxseg8ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f32mf2_m))) +void vluxseg8ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32m1))) +void vluxseg2ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32m1_m))) +void vluxseg2ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32m2))) +void vluxseg2ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32m2_m))) +void vluxseg2ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32m4))) +void vluxseg2ei64(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32m4_m))) +void vluxseg2ei64(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32mf2))) +void vluxseg2ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32mf2_m))) +void vluxseg2ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f32m1))) +void vluxseg3ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f32m1_m))) +void vluxseg3ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f32m2))) +void vluxseg3ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f32m2_m))) +void vluxseg3ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f32mf2))) +void vluxseg3ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f32mf2_m))) +void vluxseg3ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f32m1))) +void vluxseg4ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f32m1_m))) +void vluxseg4ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f32m2))) +void vluxseg4ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f32m2_m))) +void vluxseg4ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f32mf2))) +void vluxseg4ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f32mf2_m))) +void vluxseg4ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f32m1))) +void vluxseg5ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f32m1_m))) +void vluxseg5ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f32mf2))) +void vluxseg5ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f32mf2_m))) +void vluxseg5ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f32m1))) +void vluxseg6ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f32m1_m))) +void vluxseg6ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f32mf2))) +void vluxseg6ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f32mf2_m))) +void vluxseg6ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f32m1))) +void vluxseg7ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f32m1_m))) +void vluxseg7ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f32mf2))) +void vluxseg7ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f32mf2_m))) +void vluxseg7ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f32m1))) +void vluxseg8ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f32m1_m))) +void vluxseg8ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f32mf2))) +void vluxseg8ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f32mf2_m))) +void vluxseg8ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint64m1_t op18, size_t op19); + +#endif + +#if defined(__riscv_d) && defined(__riscv_zvlsseg) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f64m1))) +void vloxseg2ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f64m1_m))) +void vloxseg2ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f64m2))) +void vloxseg2ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f64m2_m))) +void vloxseg2ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f64m4))) +void vloxseg2ei8(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f64m4_m))) +void vloxseg2ei8(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f64m1))) +void vloxseg3ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f64m1_m))) +void vloxseg3ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f64m2))) +void vloxseg3ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f64m2_m))) +void vloxseg3ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f64m1))) +void vloxseg4ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f64m1_m))) +void vloxseg4ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f64m2))) +void vloxseg4ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f64m2_m))) +void vloxseg4ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f64m1))) +void vloxseg5ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f64m1_m))) +void vloxseg5ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f64m1))) +void vloxseg6ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f64m1_m))) +void vloxseg6ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f64m1))) +void vloxseg7ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f64m1_m))) +void vloxseg7ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f64m1))) +void vloxseg8ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f64m1_m))) +void vloxseg8ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f64m1))) +void vloxseg2ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f64m1_m))) +void vloxseg2ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f64m2))) +void vloxseg2ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f64m2_m))) +void vloxseg2ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f64m4))) +void vloxseg2ei16(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f64m4_m))) +void vloxseg2ei16(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f64m1))) +void vloxseg3ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f64m1_m))) +void vloxseg3ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f64m2))) +void vloxseg3ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f64m2_m))) +void vloxseg3ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f64m1))) +void vloxseg4ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f64m1_m))) +void vloxseg4ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f64m2))) +void vloxseg4ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f64m2_m))) +void vloxseg4ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f64m1))) +void vloxseg5ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f64m1_m))) +void vloxseg5ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f64m1))) +void vloxseg6ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f64m1_m))) +void vloxseg6ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f64m1))) +void vloxseg7ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f64m1_m))) +void vloxseg7ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f64m1))) +void vloxseg8ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f64m1_m))) +void vloxseg8ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f64m1))) +void vloxseg2ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f64m1_m))) +void vloxseg2ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f64m2))) +void vloxseg2ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f64m2_m))) +void vloxseg2ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f64m4))) +void vloxseg2ei32(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f64m4_m))) +void vloxseg2ei32(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f64m1))) +void vloxseg3ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f64m1_m))) +void vloxseg3ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f64m2))) +void vloxseg3ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f64m2_m))) +void vloxseg3ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f64m1))) +void vloxseg4ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f64m1_m))) +void vloxseg4ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f64m2))) +void vloxseg4ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f64m2_m))) +void vloxseg4ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f64m1))) +void vloxseg5ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f64m1_m))) +void vloxseg5ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f64m1))) +void vloxseg6ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f64m1_m))) +void vloxseg6ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f64m1))) +void vloxseg7ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f64m1_m))) +void vloxseg7ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f64m1))) +void vloxseg8ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f64m1_m))) +void vloxseg8ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f64m1))) +void vloxseg2ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f64m1_m))) +void vloxseg2ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f64m2))) +void vloxseg2ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f64m2_m))) +void vloxseg2ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f64m4))) +void vloxseg2ei64(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f64m4_m))) +void vloxseg2ei64(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f64m1))) +void vloxseg3ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f64m1_m))) +void vloxseg3ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f64m2))) +void vloxseg3ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f64m2_m))) +void vloxseg3ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f64m1))) +void vloxseg4ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f64m1_m))) +void vloxseg4ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f64m2))) +void vloxseg4ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f64m2_m))) +void vloxseg4ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f64m1))) +void vloxseg5ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f64m1_m))) +void vloxseg5ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f64m1))) +void vloxseg6ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f64m1_m))) +void vloxseg6ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f64m1))) +void vloxseg7ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f64m1_m))) +void vloxseg7ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f64m1))) +void vloxseg8ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f64m1_m))) +void vloxseg8ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_f64m1))) +void vsseg2e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_f64m1_m))) +void vsseg2e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_f64m2))) +void vsseg2e64(double * op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_f64m2_m))) +void vsseg2e64(vbool32_t op0, double * op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_f64m4))) +void vsseg2e64(double * op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_f64m4_m))) +void vsseg2e64(vbool16_t op0, double * op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_f64m1))) +void vsseg3e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_f64m1_m))) +void vsseg3e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_f64m2))) +void vsseg3e64(double * op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_f64m2_m))) +void vsseg3e64(vbool32_t op0, double * op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_f64m1))) +void vsseg4e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_f64m1_m))) +void vsseg4e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_f64m2))) +void vsseg4e64(double * op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_f64m2_m))) +void vsseg4e64(vbool32_t op0, double * op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e64_v_f64m1))) +void vsseg5e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e64_v_f64m1_m))) +void vsseg5e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e64_v_f64m1))) +void vsseg6e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e64_v_f64m1_m))) +void vsseg6e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e64_v_f64m1))) +void vsseg7e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e64_v_f64m1_m))) +void vsseg7e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e64_v_f64m1))) +void vsseg8e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e64_v_f64m1_m))) +void vsseg8e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_f64m1))) +void vssseg2e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_f64m1_m))) +void vssseg2e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_f64m2))) +void vssseg2e64(double * op0, ptrdiff_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_f64m2_m))) +void vssseg2e64(vbool32_t op0, double * op1, ptrdiff_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_f64m4))) +void vssseg2e64(double * op0, ptrdiff_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_f64m4_m))) +void vssseg2e64(vbool16_t op0, double * op1, ptrdiff_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_f64m1))) +void vssseg3e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_f64m1_m))) +void vssseg3e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_f64m2))) +void vssseg3e64(double * op0, ptrdiff_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_f64m2_m))) +void vssseg3e64(vbool32_t op0, double * op1, ptrdiff_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_f64m1))) +void vssseg4e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_f64m1_m))) +void vssseg4e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_f64m2))) +void vssseg4e64(double * op0, ptrdiff_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_f64m2_m))) +void vssseg4e64(vbool32_t op0, double * op1, ptrdiff_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e64_v_f64m1))) +void vssseg5e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e64_v_f64m1_m))) +void vssseg5e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e64_v_f64m1))) +void vssseg6e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e64_v_f64m1_m))) +void vssseg6e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e64_v_f64m1))) +void vssseg7e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e64_v_f64m1_m))) +void vssseg7e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e64_v_f64m1))) +void vssseg8e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e64_v_f64m1_m))) +void vssseg8e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f64m1))) +void vsuxseg2ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f64m1_m))) +void vsuxseg2ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f64m2))) +void vsuxseg2ei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f64m2_m))) +void vsuxseg2ei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f64m4))) +void vsuxseg2ei8(double * op0, vuint8mf2_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f64m4_m))) +void vsuxseg2ei8(vbool16_t op0, double * op1, vuint8mf2_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f64m1))) +void vsuxseg3ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f64m1_m))) +void vsuxseg3ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f64m2))) +void vsuxseg3ei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f64m2_m))) +void vsuxseg3ei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f64m1))) +void vsuxseg4ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f64m1_m))) +void vsuxseg4ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f64m2))) +void vsuxseg4ei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f64m2_m))) +void vsuxseg4ei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f64m1))) +void vsuxseg5ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f64m1_m))) +void vsuxseg5ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f64m1))) +void vsuxseg6ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f64m1_m))) +void vsuxseg6ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f64m1))) +void vsuxseg7ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f64m1_m))) +void vsuxseg7ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f64m1))) +void vsuxseg8ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f64m1_m))) +void vsuxseg8ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f64m1))) +void vsuxseg2ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f64m1_m))) +void vsuxseg2ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f64m2))) +void vsuxseg2ei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f64m2_m))) +void vsuxseg2ei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f64m4))) +void vsuxseg2ei16(double * op0, vuint16m1_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f64m4_m))) +void vsuxseg2ei16(vbool16_t op0, double * op1, vuint16m1_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f64m1))) +void vsuxseg3ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f64m1_m))) +void vsuxseg3ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f64m2))) +void vsuxseg3ei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f64m2_m))) +void vsuxseg3ei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f64m1))) +void vsuxseg4ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f64m1_m))) +void vsuxseg4ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f64m2))) +void vsuxseg4ei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f64m2_m))) +void vsuxseg4ei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f64m1))) +void vsuxseg5ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f64m1_m))) +void vsuxseg5ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f64m1))) +void vsuxseg6ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f64m1_m))) +void vsuxseg6ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f64m1))) +void vsuxseg7ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f64m1_m))) +void vsuxseg7ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f64m1))) +void vsuxseg8ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f64m1_m))) +void vsuxseg8ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f64m1))) +void vsuxseg2ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f64m1_m))) +void vsuxseg2ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f64m2))) +void vsuxseg2ei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f64m2_m))) +void vsuxseg2ei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f64m4))) +void vsuxseg2ei32(double * op0, vuint32m2_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f64m4_m))) +void vsuxseg2ei32(vbool16_t op0, double * op1, vuint32m2_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f64m1))) +void vsuxseg3ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f64m1_m))) +void vsuxseg3ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f64m2))) +void vsuxseg3ei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f64m2_m))) +void vsuxseg3ei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f64m1))) +void vsuxseg4ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f64m1_m))) +void vsuxseg4ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f64m2))) +void vsuxseg4ei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f64m2_m))) +void vsuxseg4ei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f64m1))) +void vsuxseg5ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f64m1_m))) +void vsuxseg5ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f64m1))) +void vsuxseg6ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f64m1_m))) +void vsuxseg6ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f64m1))) +void vsuxseg7ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f64m1_m))) +void vsuxseg7ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f64m1))) +void vsuxseg8ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f64m1_m))) +void vsuxseg8ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f64m1))) +void vsuxseg2ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f64m1_m))) +void vsuxseg2ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f64m2))) +void vsuxseg2ei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f64m2_m))) +void vsuxseg2ei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f64m4))) +void vsuxseg2ei64(double * op0, vuint64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f64m4_m))) +void vsuxseg2ei64(vbool16_t op0, double * op1, vuint64m4_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f64m1))) +void vsuxseg3ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f64m1_m))) +void vsuxseg3ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f64m2))) +void vsuxseg3ei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f64m2_m))) +void vsuxseg3ei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f64m1))) +void vsuxseg4ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f64m1_m))) +void vsuxseg4ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f64m2))) +void vsuxseg4ei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f64m2_m))) +void vsuxseg4ei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f64m1))) +void vsuxseg5ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f64m1_m))) +void vsuxseg5ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f64m1))) +void vsuxseg6ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f64m1_m))) +void vsuxseg6ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f64m1))) +void vsuxseg7ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f64m1_m))) +void vsuxseg7ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f64m1))) +void vsuxseg8ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f64m1_m))) +void vsuxseg8ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f64m1))) +void vsoxseg2ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f64m1_m))) +void vsoxseg2ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f64m2))) +void vsoxseg2ei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f64m2_m))) +void vsoxseg2ei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f64m4))) +void vsoxseg2ei8(double * op0, vuint8mf2_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f64m4_m))) +void vsoxseg2ei8(vbool16_t op0, double * op1, vuint8mf2_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f64m1))) +void vsoxseg3ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f64m1_m))) +void vsoxseg3ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f64m2))) +void vsoxseg3ei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f64m2_m))) +void vsoxseg3ei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f64m1))) +void vsoxseg4ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f64m1_m))) +void vsoxseg4ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f64m2))) +void vsoxseg4ei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f64m2_m))) +void vsoxseg4ei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f64m1))) +void vsoxseg5ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f64m1_m))) +void vsoxseg5ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f64m1))) +void vsoxseg6ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f64m1_m))) +void vsoxseg6ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f64m1))) +void vsoxseg7ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f64m1_m))) +void vsoxseg7ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f64m1))) +void vsoxseg8ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f64m1_m))) +void vsoxseg8ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f64m1))) +void vsoxseg2ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f64m1_m))) +void vsoxseg2ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f64m2))) +void vsoxseg2ei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f64m2_m))) +void vsoxseg2ei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f64m4))) +void vsoxseg2ei16(double * op0, vuint16m1_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f64m4_m))) +void vsoxseg2ei16(vbool16_t op0, double * op1, vuint16m1_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f64m1))) +void vsoxseg3ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f64m1_m))) +void vsoxseg3ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f64m2))) +void vsoxseg3ei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f64m2_m))) +void vsoxseg3ei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f64m1))) +void vsoxseg4ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f64m1_m))) +void vsoxseg4ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f64m2))) +void vsoxseg4ei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f64m2_m))) +void vsoxseg4ei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f64m1))) +void vsoxseg5ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f64m1_m))) +void vsoxseg5ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f64m1))) +void vsoxseg6ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f64m1_m))) +void vsoxseg6ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f64m1))) +void vsoxseg7ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f64m1_m))) +void vsoxseg7ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f64m1))) +void vsoxseg8ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f64m1_m))) +void vsoxseg8ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f64m1))) +void vsoxseg2ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f64m1_m))) +void vsoxseg2ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f64m2))) +void vsoxseg2ei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f64m2_m))) +void vsoxseg2ei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f64m4))) +void vsoxseg2ei32(double * op0, vuint32m2_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f64m4_m))) +void vsoxseg2ei32(vbool16_t op0, double * op1, vuint32m2_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f64m1))) +void vsoxseg3ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f64m1_m))) +void vsoxseg3ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f64m2))) +void vsoxseg3ei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f64m2_m))) +void vsoxseg3ei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f64m1))) +void vsoxseg4ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f64m1_m))) +void vsoxseg4ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f64m2))) +void vsoxseg4ei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f64m2_m))) +void vsoxseg4ei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f64m1))) +void vsoxseg5ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f64m1_m))) +void vsoxseg5ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f64m1))) +void vsoxseg6ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f64m1_m))) +void vsoxseg6ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f64m1))) +void vsoxseg7ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f64m1_m))) +void vsoxseg7ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f64m1))) +void vsoxseg8ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f64m1_m))) +void vsoxseg8ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f64m1))) +void vsoxseg2ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f64m1_m))) +void vsoxseg2ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f64m2))) +void vsoxseg2ei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f64m2_m))) +void vsoxseg2ei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f64m4))) +void vsoxseg2ei64(double * op0, vuint64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f64m4_m))) +void vsoxseg2ei64(vbool16_t op0, double * op1, vuint64m4_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f64m1))) +void vsoxseg3ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f64m1_m))) +void vsoxseg3ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f64m2))) +void vsoxseg3ei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f64m2_m))) +void vsoxseg3ei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f64m1))) +void vsoxseg4ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f64m1_m))) +void vsoxseg4ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f64m2))) +void vsoxseg4ei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f64m2_m))) +void vsoxseg4ei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f64m1))) +void vsoxseg5ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f64m1_m))) +void vsoxseg5ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f64m1))) +void vsoxseg6ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f64m1_m))) +void vsoxseg6ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f64m1))) +void vsoxseg7ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f64m1_m))) +void vsoxseg7ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f64m1))) +void vsoxseg8ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f64m1_m))) +void vsoxseg8ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_f64m1_m))) +void vlseg2e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_f64m2_m))) +void vlseg2e64(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_f64m4_m))) +void vlseg2e64(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64_v_f64m1_m))) +void vlseg3e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64_v_f64m2_m))) +void vlseg3e64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64_v_f64m1_m))) +void vlseg4e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64_v_f64m2_m))) +void vlseg4e64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e64_v_f64m1_m))) +void vlseg5e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e64_v_f64m1_m))) +void vlseg6e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e64_v_f64m1_m))) +void vlseg7e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e64_v_f64m1_m))) +void vlseg8e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_f64m1_m))) +void vlseg2e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_f64m2_m))) +void vlseg2e64ff(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_f64m4_m))) +void vlseg2e64ff(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64ff_v_f64m1_m))) +void vlseg3e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64ff_v_f64m2_m))) +void vlseg3e64ff(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64ff_v_f64m1_m))) +void vlseg4e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64ff_v_f64m2_m))) +void vlseg4e64ff(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e64ff_v_f64m1_m))) +void vlseg5e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e64ff_v_f64m1_m))) +void vlseg6e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e64ff_v_f64m1_m))) +void vlseg7e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e64ff_v_f64m1_m))) +void vlseg8e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_f64m1_m))) +void vlsseg2e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_f64m2_m))) +void vlsseg2e64(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_f64m4_m))) +void vlsseg2e64(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e64_v_f64m1_m))) +void vlsseg3e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e64_v_f64m2_m))) +void vlsseg3e64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e64_v_f64m1_m))) +void vlsseg4e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e64_v_f64m2_m))) +void vlsseg4e64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e64_v_f64m1_m))) +void vlsseg5e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e64_v_f64m1_m))) +void vlsseg6e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e64_v_f64m1_m))) +void vlsseg7e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e64_v_f64m1_m))) +void vlsseg8e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f64m1))) +void vluxseg2ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f64m1_m))) +void vluxseg2ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f64m2))) +void vluxseg2ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f64m2_m))) +void vluxseg2ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f64m4))) +void vluxseg2ei8(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f64m4_m))) +void vluxseg2ei8(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f64m1))) +void vluxseg3ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f64m1_m))) +void vluxseg3ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f64m2))) +void vluxseg3ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f64m2_m))) +void vluxseg3ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f64m1))) +void vluxseg4ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f64m1_m))) +void vluxseg4ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f64m2))) +void vluxseg4ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f64m2_m))) +void vluxseg4ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f64m1))) +void vluxseg5ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f64m1_m))) +void vluxseg5ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f64m1))) +void vluxseg6ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f64m1_m))) +void vluxseg6ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f64m1))) +void vluxseg7ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f64m1_m))) +void vluxseg7ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f64m1))) +void vluxseg8ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f64m1_m))) +void vluxseg8ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f64m1))) +void vluxseg2ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f64m1_m))) +void vluxseg2ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f64m2))) +void vluxseg2ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f64m2_m))) +void vluxseg2ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f64m4))) +void vluxseg2ei16(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f64m4_m))) +void vluxseg2ei16(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f64m1))) +void vluxseg3ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f64m1_m))) +void vluxseg3ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f64m2))) +void vluxseg3ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f64m2_m))) +void vluxseg3ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f64m1))) +void vluxseg4ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f64m1_m))) +void vluxseg4ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f64m2))) +void vluxseg4ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f64m2_m))) +void vluxseg4ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f64m1))) +void vluxseg5ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f64m1_m))) +void vluxseg5ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f64m1))) +void vluxseg6ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f64m1_m))) +void vluxseg6ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f64m1))) +void vluxseg7ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f64m1_m))) +void vluxseg7ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f64m1))) +void vluxseg8ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f64m1_m))) +void vluxseg8ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f64m1))) +void vluxseg2ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f64m1_m))) +void vluxseg2ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f64m2))) +void vluxseg2ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f64m2_m))) +void vluxseg2ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f64m4))) +void vluxseg2ei32(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f64m4_m))) +void vluxseg2ei32(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f64m1))) +void vluxseg3ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f64m1_m))) +void vluxseg3ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f64m2))) +void vluxseg3ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f64m2_m))) +void vluxseg3ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f64m1))) +void vluxseg4ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f64m1_m))) +void vluxseg4ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f64m2))) +void vluxseg4ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f64m2_m))) +void vluxseg4ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f64m1))) +void vluxseg5ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f64m1_m))) +void vluxseg5ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f64m1))) +void vluxseg6ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f64m1_m))) +void vluxseg6ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f64m1))) +void vluxseg7ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f64m1_m))) +void vluxseg7ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f64m1))) +void vluxseg8ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f64m1_m))) +void vluxseg8ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f64m1))) +void vluxseg2ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f64m1_m))) +void vluxseg2ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f64m2))) +void vluxseg2ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f64m2_m))) +void vluxseg2ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f64m4))) +void vluxseg2ei64(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f64m4_m))) +void vluxseg2ei64(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f64m1))) +void vluxseg3ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f64m1_m))) +void vluxseg3ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f64m2))) +void vluxseg3ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f64m2_m))) +void vluxseg3ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f64m1))) +void vluxseg4ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f64m1_m))) +void vluxseg4ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f64m2))) +void vluxseg4ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f64m2_m))) +void vluxseg4ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f64m1))) +void vluxseg5ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f64m1_m))) +void vluxseg5ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f64m1))) +void vluxseg6ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f64m1_m))) +void vluxseg6ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f64m1))) +void vluxseg7ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f64m1_m))) +void vluxseg7ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f64m1))) +void vluxseg8ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f64m1_m))) +void vluxseg8ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint64m1_t op18, size_t op19); + +#endif + +#if defined(__riscv_zfh) && defined(__riscv_zvlsseg) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16m1))) +void vloxseg2ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16m1_m))) +void vloxseg2ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16m2))) +void vloxseg2ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16m2_m))) +void vloxseg2ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16m4))) +void vloxseg2ei8(vfloat16m4_t * op0, vfloat16m4_t * op1, const _Float16 * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16m4_m))) +void vloxseg2ei8(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16mf2))) +void vloxseg2ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16mf2_m))) +void vloxseg2ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16mf4))) +void vloxseg2ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16mf4_m))) +void vloxseg2ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16m1))) +void vloxseg3ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16m1_m))) +void vloxseg3ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16m2))) +void vloxseg3ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16m2_m))) +void vloxseg3ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16mf2))) +void vloxseg3ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16mf2_m))) +void vloxseg3ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16mf4))) +void vloxseg3ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16mf4_m))) +void vloxseg3ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16m1))) +void vloxseg4ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16m1_m))) +void vloxseg4ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16m2))) +void vloxseg4ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16m2_m))) +void vloxseg4ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16mf2))) +void vloxseg4ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16mf2_m))) +void vloxseg4ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16mf4))) +void vloxseg4ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16mf4_m))) +void vloxseg4ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f16m1))) +void vloxseg5ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f16m1_m))) +void vloxseg5ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f16mf2))) +void vloxseg5ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f16mf2_m))) +void vloxseg5ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f16mf4))) +void vloxseg5ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f16mf4_m))) +void vloxseg5ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f16m1))) +void vloxseg6ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f16m1_m))) +void vloxseg6ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f16mf2))) +void vloxseg6ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f16mf2_m))) +void vloxseg6ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f16mf4))) +void vloxseg6ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f16mf4_m))) +void vloxseg6ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f16m1))) +void vloxseg7ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f16m1_m))) +void vloxseg7ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f16mf2))) +void vloxseg7ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f16mf2_m))) +void vloxseg7ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f16mf4))) +void vloxseg7ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f16mf4_m))) +void vloxseg7ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f16m1))) +void vloxseg8ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f16m1_m))) +void vloxseg8ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f16mf2))) +void vloxseg8ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f16mf2_m))) +void vloxseg8ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f16mf4))) +void vloxseg8ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f16mf4_m))) +void vloxseg8ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16m1))) +void vloxseg2ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16m1_m))) +void vloxseg2ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16m2))) +void vloxseg2ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16m2_m))) +void vloxseg2ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16m4))) +void vloxseg2ei16(vfloat16m4_t * op0, vfloat16m4_t * op1, const _Float16 * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16m4_m))) +void vloxseg2ei16(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16mf2))) +void vloxseg2ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16mf2_m))) +void vloxseg2ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16mf4))) +void vloxseg2ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16mf4_m))) +void vloxseg2ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16m1))) +void vloxseg3ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16m1_m))) +void vloxseg3ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16m2))) +void vloxseg3ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16m2_m))) +void vloxseg3ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16mf2))) +void vloxseg3ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16mf2_m))) +void vloxseg3ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16mf4))) +void vloxseg3ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16mf4_m))) +void vloxseg3ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16m1))) +void vloxseg4ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16m1_m))) +void vloxseg4ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16m2))) +void vloxseg4ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16m2_m))) +void vloxseg4ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16mf2))) +void vloxseg4ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16mf2_m))) +void vloxseg4ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16mf4))) +void vloxseg4ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16mf4_m))) +void vloxseg4ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f16m1))) +void vloxseg5ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f16m1_m))) +void vloxseg5ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f16mf2))) +void vloxseg5ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f16mf2_m))) +void vloxseg5ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f16mf4))) +void vloxseg5ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f16mf4_m))) +void vloxseg5ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f16m1))) +void vloxseg6ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f16m1_m))) +void vloxseg6ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f16mf2))) +void vloxseg6ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f16mf2_m))) +void vloxseg6ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f16mf4))) +void vloxseg6ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f16mf4_m))) +void vloxseg6ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f16m1))) +void vloxseg7ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f16m1_m))) +void vloxseg7ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f16mf2))) +void vloxseg7ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f16mf2_m))) +void vloxseg7ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f16mf4))) +void vloxseg7ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f16mf4_m))) +void vloxseg7ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f16m1))) +void vloxseg8ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f16m1_m))) +void vloxseg8ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f16mf2))) +void vloxseg8ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f16mf2_m))) +void vloxseg8ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f16mf4))) +void vloxseg8ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f16mf4_m))) +void vloxseg8ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16m1))) +void vloxseg2ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16m1_m))) +void vloxseg2ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16m2))) +void vloxseg2ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16m2_m))) +void vloxseg2ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16m4))) +void vloxseg2ei32(vfloat16m4_t * op0, vfloat16m4_t * op1, const _Float16 * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16m4_m))) +void vloxseg2ei32(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16mf2))) +void vloxseg2ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16mf2_m))) +void vloxseg2ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16mf4))) +void vloxseg2ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16mf4_m))) +void vloxseg2ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16m1))) +void vloxseg3ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16m1_m))) +void vloxseg3ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16m2))) +void vloxseg3ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16m2_m))) +void vloxseg3ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16mf2))) +void vloxseg3ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16mf2_m))) +void vloxseg3ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16mf4))) +void vloxseg3ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16mf4_m))) +void vloxseg3ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16m1))) +void vloxseg4ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16m1_m))) +void vloxseg4ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16m2))) +void vloxseg4ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16m2_m))) +void vloxseg4ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16mf2))) +void vloxseg4ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16mf2_m))) +void vloxseg4ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16mf4))) +void vloxseg4ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16mf4_m))) +void vloxseg4ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f16m1))) +void vloxseg5ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f16m1_m))) +void vloxseg5ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f16mf2))) +void vloxseg5ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f16mf2_m))) +void vloxseg5ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f16mf4))) +void vloxseg5ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f16mf4_m))) +void vloxseg5ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f16m1))) +void vloxseg6ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f16m1_m))) +void vloxseg6ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f16mf2))) +void vloxseg6ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f16mf2_m))) +void vloxseg6ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f16mf4))) +void vloxseg6ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f16mf4_m))) +void vloxseg6ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f16m1))) +void vloxseg7ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f16m1_m))) +void vloxseg7ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f16mf2))) +void vloxseg7ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f16mf2_m))) +void vloxseg7ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f16mf4))) +void vloxseg7ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f16mf4_m))) +void vloxseg7ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f16m1))) +void vloxseg8ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f16m1_m))) +void vloxseg8ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f16mf2))) +void vloxseg8ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f16mf2_m))) +void vloxseg8ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f16mf4))) +void vloxseg8ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f16mf4_m))) +void vloxseg8ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16m1))) +void vloxseg2ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16m1_m))) +void vloxseg2ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16m2))) +void vloxseg2ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16m2_m))) +void vloxseg2ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16mf2))) +void vloxseg2ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16mf2_m))) +void vloxseg2ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16mf4))) +void vloxseg2ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16mf4_m))) +void vloxseg2ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16m1))) +void vloxseg3ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16m1_m))) +void vloxseg3ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16m2))) +void vloxseg3ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16m2_m))) +void vloxseg3ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16mf2))) +void vloxseg3ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16mf2_m))) +void vloxseg3ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16mf4))) +void vloxseg3ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16mf4_m))) +void vloxseg3ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16m1))) +void vloxseg4ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16m1_m))) +void vloxseg4ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16m2))) +void vloxseg4ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16m2_m))) +void vloxseg4ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16mf2))) +void vloxseg4ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16mf2_m))) +void vloxseg4ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16mf4))) +void vloxseg4ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16mf4_m))) +void vloxseg4ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f16m1))) +void vloxseg5ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f16m1_m))) +void vloxseg5ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f16mf2))) +void vloxseg5ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f16mf2_m))) +void vloxseg5ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f16mf4))) +void vloxseg5ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f16mf4_m))) +void vloxseg5ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f16m1))) +void vloxseg6ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f16m1_m))) +void vloxseg6ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f16mf2))) +void vloxseg6ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f16mf2_m))) +void vloxseg6ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f16mf4))) +void vloxseg6ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f16mf4_m))) +void vloxseg6ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f16m1))) +void vloxseg7ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f16m1_m))) +void vloxseg7ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f16mf2))) +void vloxseg7ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f16mf2_m))) +void vloxseg7ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f16mf4))) +void vloxseg7ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f16mf4_m))) +void vloxseg7ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f16m1))) +void vloxseg8ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f16m1_m))) +void vloxseg8ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f16mf2))) +void vloxseg8ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f16mf2_m))) +void vloxseg8ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f16mf4))) +void vloxseg8ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f16mf4_m))) +void vloxseg8ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16m1))) +void vsseg2e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16m1_m))) +void vsseg2e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16m2))) +void vsseg2e16(_Float16 * op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16m2_m))) +void vsseg2e16(vbool8_t op0, _Float16 * op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16m4))) +void vsseg2e16(_Float16 * op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16m4_m))) +void vsseg2e16(vbool4_t op0, _Float16 * op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16mf2))) +void vsseg2e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16mf2_m))) +void vsseg2e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16mf4))) +void vsseg2e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16mf4_m))) +void vsseg2e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16m1))) +void vsseg3e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16m1_m))) +void vsseg3e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16m2))) +void vsseg3e16(_Float16 * op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16m2_m))) +void vsseg3e16(vbool8_t op0, _Float16 * op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16mf2))) +void vsseg3e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16mf2_m))) +void vsseg3e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16mf4))) +void vsseg3e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16mf4_m))) +void vsseg3e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16m1))) +void vsseg4e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16m1_m))) +void vsseg4e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16m2))) +void vsseg4e16(_Float16 * op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16m2_m))) +void vsseg4e16(vbool8_t op0, _Float16 * op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16mf2))) +void vsseg4e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16mf2_m))) +void vsseg4e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16mf4))) +void vsseg4e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16mf4_m))) +void vsseg4e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_f16m1))) +void vsseg5e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_f16m1_m))) +void vsseg5e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_f16mf2))) +void vsseg5e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_f16mf2_m))) +void vsseg5e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_f16mf4))) +void vsseg5e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_f16mf4_m))) +void vsseg5e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_f16m1))) +void vsseg6e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_f16m1_m))) +void vsseg6e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_f16mf2))) +void vsseg6e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_f16mf2_m))) +void vsseg6e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_f16mf4))) +void vsseg6e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_f16mf4_m))) +void vsseg6e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_f16m1))) +void vsseg7e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_f16m1_m))) +void vsseg7e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_f16mf2))) +void vsseg7e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_f16mf2_m))) +void vsseg7e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_f16mf4))) +void vsseg7e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_f16mf4_m))) +void vsseg7e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_f16m1))) +void vsseg8e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_f16m1_m))) +void vsseg8e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_f16mf2))) +void vsseg8e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_f16mf2_m))) +void vsseg8e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_f16mf4))) +void vsseg8e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_f16mf4_m))) +void vsseg8e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16m1))) +void vssseg2e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16m1_m))) +void vssseg2e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16m2))) +void vssseg2e16(_Float16 * op0, ptrdiff_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16m2_m))) +void vssseg2e16(vbool8_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16m4))) +void vssseg2e16(_Float16 * op0, ptrdiff_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16m4_m))) +void vssseg2e16(vbool4_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16mf2))) +void vssseg2e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16mf2_m))) +void vssseg2e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16mf4))) +void vssseg2e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16mf4_m))) +void vssseg2e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16m1))) +void vssseg3e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16m1_m))) +void vssseg3e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16m2))) +void vssseg3e16(_Float16 * op0, ptrdiff_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16m2_m))) +void vssseg3e16(vbool8_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16mf2))) +void vssseg3e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16mf2_m))) +void vssseg3e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16mf4))) +void vssseg3e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16mf4_m))) +void vssseg3e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16m1))) +void vssseg4e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16m1_m))) +void vssseg4e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16m2))) +void vssseg4e16(_Float16 * op0, ptrdiff_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16m2_m))) +void vssseg4e16(vbool8_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16mf2))) +void vssseg4e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16mf2_m))) +void vssseg4e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16mf4))) +void vssseg4e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16mf4_m))) +void vssseg4e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_f16m1))) +void vssseg5e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_f16m1_m))) +void vssseg5e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_f16mf2))) +void vssseg5e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_f16mf2_m))) +void vssseg5e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_f16mf4))) +void vssseg5e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_f16mf4_m))) +void vssseg5e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_f16m1))) +void vssseg6e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_f16m1_m))) +void vssseg6e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_f16mf2))) +void vssseg6e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_f16mf2_m))) +void vssseg6e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_f16mf4))) +void vssseg6e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_f16mf4_m))) +void vssseg6e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_f16m1))) +void vssseg7e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_f16m1_m))) +void vssseg7e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_f16mf2))) +void vssseg7e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_f16mf2_m))) +void vssseg7e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_f16mf4))) +void vssseg7e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_f16mf4_m))) +void vssseg7e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_f16m1))) +void vssseg8e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_f16m1_m))) +void vssseg8e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_f16mf2))) +void vssseg8e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_f16mf2_m))) +void vssseg8e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_f16mf4))) +void vssseg8e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_f16mf4_m))) +void vssseg8e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16m1))) +void vsuxseg2ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16m1_m))) +void vsuxseg2ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16m2))) +void vsuxseg2ei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16m2_m))) +void vsuxseg2ei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16m4))) +void vsuxseg2ei8(_Float16 * op0, vuint8m2_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16m4_m))) +void vsuxseg2ei8(vbool4_t op0, _Float16 * op1, vuint8m2_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16mf2))) +void vsuxseg2ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16mf2_m))) +void vsuxseg2ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16mf4))) +void vsuxseg2ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16mf4_m))) +void vsuxseg2ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16m1))) +void vsuxseg3ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16m1_m))) +void vsuxseg3ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16m2))) +void vsuxseg3ei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16m2_m))) +void vsuxseg3ei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16mf2))) +void vsuxseg3ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16mf2_m))) +void vsuxseg3ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16mf4))) +void vsuxseg3ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16mf4_m))) +void vsuxseg3ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16m1))) +void vsuxseg4ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16m1_m))) +void vsuxseg4ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16m2))) +void vsuxseg4ei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16m2_m))) +void vsuxseg4ei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16mf2))) +void vsuxseg4ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16mf2_m))) +void vsuxseg4ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16mf4))) +void vsuxseg4ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16mf4_m))) +void vsuxseg4ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f16m1))) +void vsuxseg5ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f16m1_m))) +void vsuxseg5ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f16mf2))) +void vsuxseg5ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f16mf2_m))) +void vsuxseg5ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f16mf4))) +void vsuxseg5ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f16mf4_m))) +void vsuxseg5ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f16m1))) +void vsuxseg6ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f16m1_m))) +void vsuxseg6ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f16mf2))) +void vsuxseg6ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f16mf2_m))) +void vsuxseg6ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f16mf4))) +void vsuxseg6ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f16mf4_m))) +void vsuxseg6ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f16m1))) +void vsuxseg7ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f16m1_m))) +void vsuxseg7ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f16mf2))) +void vsuxseg7ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f16mf2_m))) +void vsuxseg7ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f16mf4))) +void vsuxseg7ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f16mf4_m))) +void vsuxseg7ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f16m1))) +void vsuxseg8ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f16m1_m))) +void vsuxseg8ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f16mf2))) +void vsuxseg8ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f16mf2_m))) +void vsuxseg8ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f16mf4))) +void vsuxseg8ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f16mf4_m))) +void vsuxseg8ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16m1))) +void vsuxseg2ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16m1_m))) +void vsuxseg2ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16m2))) +void vsuxseg2ei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16m2_m))) +void vsuxseg2ei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16m4))) +void vsuxseg2ei16(_Float16 * op0, vuint16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16m4_m))) +void vsuxseg2ei16(vbool4_t op0, _Float16 * op1, vuint16m4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16mf2))) +void vsuxseg2ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16mf2_m))) +void vsuxseg2ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16mf4))) +void vsuxseg2ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16mf4_m))) +void vsuxseg2ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16m1))) +void vsuxseg3ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16m1_m))) +void vsuxseg3ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16m2))) +void vsuxseg3ei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16m2_m))) +void vsuxseg3ei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16mf2))) +void vsuxseg3ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16mf2_m))) +void vsuxseg3ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16mf4))) +void vsuxseg3ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16mf4_m))) +void vsuxseg3ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16m1))) +void vsuxseg4ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16m1_m))) +void vsuxseg4ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16m2))) +void vsuxseg4ei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16m2_m))) +void vsuxseg4ei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16mf2))) +void vsuxseg4ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16mf2_m))) +void vsuxseg4ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16mf4))) +void vsuxseg4ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16mf4_m))) +void vsuxseg4ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f16m1))) +void vsuxseg5ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f16m1_m))) +void vsuxseg5ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f16mf2))) +void vsuxseg5ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f16mf2_m))) +void vsuxseg5ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f16mf4))) +void vsuxseg5ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f16mf4_m))) +void vsuxseg5ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f16m1))) +void vsuxseg6ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f16m1_m))) +void vsuxseg6ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f16mf2))) +void vsuxseg6ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f16mf2_m))) +void vsuxseg6ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f16mf4))) +void vsuxseg6ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f16mf4_m))) +void vsuxseg6ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f16m1))) +void vsuxseg7ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f16m1_m))) +void vsuxseg7ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f16mf2))) +void vsuxseg7ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f16mf2_m))) +void vsuxseg7ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f16mf4))) +void vsuxseg7ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f16mf4_m))) +void vsuxseg7ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f16m1))) +void vsuxseg8ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f16m1_m))) +void vsuxseg8ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f16mf2))) +void vsuxseg8ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f16mf2_m))) +void vsuxseg8ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f16mf4))) +void vsuxseg8ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f16mf4_m))) +void vsuxseg8ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16m1))) +void vsuxseg2ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16m1_m))) +void vsuxseg2ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16m2))) +void vsuxseg2ei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16m2_m))) +void vsuxseg2ei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16m4))) +void vsuxseg2ei32(_Float16 * op0, vuint32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16m4_m))) +void vsuxseg2ei32(vbool4_t op0, _Float16 * op1, vuint32m8_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16mf2))) +void vsuxseg2ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16mf2_m))) +void vsuxseg2ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16mf4))) +void vsuxseg2ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16mf4_m))) +void vsuxseg2ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16m1))) +void vsuxseg3ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16m1_m))) +void vsuxseg3ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16m2))) +void vsuxseg3ei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16m2_m))) +void vsuxseg3ei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16mf2))) +void vsuxseg3ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16mf2_m))) +void vsuxseg3ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16mf4))) +void vsuxseg3ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16mf4_m))) +void vsuxseg3ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16m1))) +void vsuxseg4ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16m1_m))) +void vsuxseg4ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16m2))) +void vsuxseg4ei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16m2_m))) +void vsuxseg4ei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16mf2))) +void vsuxseg4ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16mf2_m))) +void vsuxseg4ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16mf4))) +void vsuxseg4ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16mf4_m))) +void vsuxseg4ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f16m1))) +void vsuxseg5ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f16m1_m))) +void vsuxseg5ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f16mf2))) +void vsuxseg5ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f16mf2_m))) +void vsuxseg5ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f16mf4))) +void vsuxseg5ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f16mf4_m))) +void vsuxseg5ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f16m1))) +void vsuxseg6ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f16m1_m))) +void vsuxseg6ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f16mf2))) +void vsuxseg6ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f16mf2_m))) +void vsuxseg6ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f16mf4))) +void vsuxseg6ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f16mf4_m))) +void vsuxseg6ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f16m1))) +void vsuxseg7ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f16m1_m))) +void vsuxseg7ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f16mf2))) +void vsuxseg7ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f16mf2_m))) +void vsuxseg7ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f16mf4))) +void vsuxseg7ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f16mf4_m))) +void vsuxseg7ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f16m1))) +void vsuxseg8ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f16m1_m))) +void vsuxseg8ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f16mf2))) +void vsuxseg8ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f16mf2_m))) +void vsuxseg8ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f16mf4))) +void vsuxseg8ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f16mf4_m))) +void vsuxseg8ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16m1))) +void vsuxseg2ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16m1_m))) +void vsuxseg2ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16m2))) +void vsuxseg2ei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16m2_m))) +void vsuxseg2ei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16mf2))) +void vsuxseg2ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16mf2_m))) +void vsuxseg2ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16mf4))) +void vsuxseg2ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16mf4_m))) +void vsuxseg2ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16m1))) +void vsuxseg3ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16m1_m))) +void vsuxseg3ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16m2))) +void vsuxseg3ei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16m2_m))) +void vsuxseg3ei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16mf2))) +void vsuxseg3ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16mf2_m))) +void vsuxseg3ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16mf4))) +void vsuxseg3ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16mf4_m))) +void vsuxseg3ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16m1))) +void vsuxseg4ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16m1_m))) +void vsuxseg4ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16m2))) +void vsuxseg4ei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16m2_m))) +void vsuxseg4ei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16mf2))) +void vsuxseg4ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16mf2_m))) +void vsuxseg4ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16mf4))) +void vsuxseg4ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16mf4_m))) +void vsuxseg4ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f16m1))) +void vsuxseg5ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f16m1_m))) +void vsuxseg5ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f16mf2))) +void vsuxseg5ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f16mf2_m))) +void vsuxseg5ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f16mf4))) +void vsuxseg5ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f16mf4_m))) +void vsuxseg5ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f16m1))) +void vsuxseg6ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f16m1_m))) +void vsuxseg6ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f16mf2))) +void vsuxseg6ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f16mf2_m))) +void vsuxseg6ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f16mf4))) +void vsuxseg6ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f16mf4_m))) +void vsuxseg6ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f16m1))) +void vsuxseg7ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f16m1_m))) +void vsuxseg7ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f16mf2))) +void vsuxseg7ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f16mf2_m))) +void vsuxseg7ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f16mf4))) +void vsuxseg7ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f16mf4_m))) +void vsuxseg7ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f16m1))) +void vsuxseg8ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f16m1_m))) +void vsuxseg8ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f16mf2))) +void vsuxseg8ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f16mf2_m))) +void vsuxseg8ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f16mf4))) +void vsuxseg8ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f16mf4_m))) +void vsuxseg8ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16m1))) +void vsoxseg2ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16m1_m))) +void vsoxseg2ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16m2))) +void vsoxseg2ei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16m2_m))) +void vsoxseg2ei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16m4))) +void vsoxseg2ei8(_Float16 * op0, vuint8m2_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16m4_m))) +void vsoxseg2ei8(vbool4_t op0, _Float16 * op1, vuint8m2_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16mf2))) +void vsoxseg2ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16mf2_m))) +void vsoxseg2ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16mf4))) +void vsoxseg2ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16mf4_m))) +void vsoxseg2ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16m1))) +void vsoxseg3ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16m1_m))) +void vsoxseg3ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16m2))) +void vsoxseg3ei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16m2_m))) +void vsoxseg3ei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16mf2))) +void vsoxseg3ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16mf2_m))) +void vsoxseg3ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16mf4))) +void vsoxseg3ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16mf4_m))) +void vsoxseg3ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16m1))) +void vsoxseg4ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16m1_m))) +void vsoxseg4ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16m2))) +void vsoxseg4ei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16m2_m))) +void vsoxseg4ei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16mf2))) +void vsoxseg4ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16mf2_m))) +void vsoxseg4ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16mf4))) +void vsoxseg4ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16mf4_m))) +void vsoxseg4ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f16m1))) +void vsoxseg5ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f16m1_m))) +void vsoxseg5ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f16mf2))) +void vsoxseg5ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f16mf2_m))) +void vsoxseg5ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f16mf4))) +void vsoxseg5ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f16mf4_m))) +void vsoxseg5ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f16m1))) +void vsoxseg6ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f16m1_m))) +void vsoxseg6ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f16mf2))) +void vsoxseg6ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f16mf2_m))) +void vsoxseg6ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f16mf4))) +void vsoxseg6ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f16mf4_m))) +void vsoxseg6ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f16m1))) +void vsoxseg7ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f16m1_m))) +void vsoxseg7ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f16mf2))) +void vsoxseg7ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f16mf2_m))) +void vsoxseg7ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f16mf4))) +void vsoxseg7ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f16mf4_m))) +void vsoxseg7ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f16m1))) +void vsoxseg8ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f16m1_m))) +void vsoxseg8ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f16mf2))) +void vsoxseg8ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f16mf2_m))) +void vsoxseg8ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f16mf4))) +void vsoxseg8ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f16mf4_m))) +void vsoxseg8ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16m1))) +void vsoxseg2ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16m1_m))) +void vsoxseg2ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16m2))) +void vsoxseg2ei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16m2_m))) +void vsoxseg2ei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16m4))) +void vsoxseg2ei16(_Float16 * op0, vuint16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16m4_m))) +void vsoxseg2ei16(vbool4_t op0, _Float16 * op1, vuint16m4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16mf2))) +void vsoxseg2ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16mf2_m))) +void vsoxseg2ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16mf4))) +void vsoxseg2ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16mf4_m))) +void vsoxseg2ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16m1))) +void vsoxseg3ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16m1_m))) +void vsoxseg3ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16m2))) +void vsoxseg3ei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16m2_m))) +void vsoxseg3ei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16mf2))) +void vsoxseg3ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16mf2_m))) +void vsoxseg3ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16mf4))) +void vsoxseg3ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16mf4_m))) +void vsoxseg3ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16m1))) +void vsoxseg4ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16m1_m))) +void vsoxseg4ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16m2))) +void vsoxseg4ei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16m2_m))) +void vsoxseg4ei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16mf2))) +void vsoxseg4ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16mf2_m))) +void vsoxseg4ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16mf4))) +void vsoxseg4ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16mf4_m))) +void vsoxseg4ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f16m1))) +void vsoxseg5ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f16m1_m))) +void vsoxseg5ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f16mf2))) +void vsoxseg5ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f16mf2_m))) +void vsoxseg5ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f16mf4))) +void vsoxseg5ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f16mf4_m))) +void vsoxseg5ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f16m1))) +void vsoxseg6ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f16m1_m))) +void vsoxseg6ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f16mf2))) +void vsoxseg6ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f16mf2_m))) +void vsoxseg6ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f16mf4))) +void vsoxseg6ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f16mf4_m))) +void vsoxseg6ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f16m1))) +void vsoxseg7ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f16m1_m))) +void vsoxseg7ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f16mf2))) +void vsoxseg7ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f16mf2_m))) +void vsoxseg7ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f16mf4))) +void vsoxseg7ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f16mf4_m))) +void vsoxseg7ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f16m1))) +void vsoxseg8ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f16m1_m))) +void vsoxseg8ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f16mf2))) +void vsoxseg8ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f16mf2_m))) +void vsoxseg8ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f16mf4))) +void vsoxseg8ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f16mf4_m))) +void vsoxseg8ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16m1))) +void vsoxseg2ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16m1_m))) +void vsoxseg2ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16m2))) +void vsoxseg2ei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16m2_m))) +void vsoxseg2ei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16m4))) +void vsoxseg2ei32(_Float16 * op0, vuint32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16m4_m))) +void vsoxseg2ei32(vbool4_t op0, _Float16 * op1, vuint32m8_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16mf2))) +void vsoxseg2ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16mf2_m))) +void vsoxseg2ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16mf4))) +void vsoxseg2ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16mf4_m))) +void vsoxseg2ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16m1))) +void vsoxseg3ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16m1_m))) +void vsoxseg3ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16m2))) +void vsoxseg3ei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16m2_m))) +void vsoxseg3ei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16mf2))) +void vsoxseg3ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16mf2_m))) +void vsoxseg3ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16mf4))) +void vsoxseg3ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16mf4_m))) +void vsoxseg3ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16m1))) +void vsoxseg4ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16m1_m))) +void vsoxseg4ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16m2))) +void vsoxseg4ei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16m2_m))) +void vsoxseg4ei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16mf2))) +void vsoxseg4ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16mf2_m))) +void vsoxseg4ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16mf4))) +void vsoxseg4ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16mf4_m))) +void vsoxseg4ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f16m1))) +void vsoxseg5ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f16m1_m))) +void vsoxseg5ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f16mf2))) +void vsoxseg5ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f16mf2_m))) +void vsoxseg5ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f16mf4))) +void vsoxseg5ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f16mf4_m))) +void vsoxseg5ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f16m1))) +void vsoxseg6ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f16m1_m))) +void vsoxseg6ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f16mf2))) +void vsoxseg6ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f16mf2_m))) +void vsoxseg6ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f16mf4))) +void vsoxseg6ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f16mf4_m))) +void vsoxseg6ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f16m1))) +void vsoxseg7ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f16m1_m))) +void vsoxseg7ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f16mf2))) +void vsoxseg7ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f16mf2_m))) +void vsoxseg7ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f16mf4))) +void vsoxseg7ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f16mf4_m))) +void vsoxseg7ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f16m1))) +void vsoxseg8ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f16m1_m))) +void vsoxseg8ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f16mf2))) +void vsoxseg8ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f16mf2_m))) +void vsoxseg8ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f16mf4))) +void vsoxseg8ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f16mf4_m))) +void vsoxseg8ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16m1))) +void vsoxseg2ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16m1_m))) +void vsoxseg2ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16m2))) +void vsoxseg2ei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16m2_m))) +void vsoxseg2ei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16mf2))) +void vsoxseg2ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16mf2_m))) +void vsoxseg2ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16mf4))) +void vsoxseg2ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16mf4_m))) +void vsoxseg2ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16m1))) +void vsoxseg3ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16m1_m))) +void vsoxseg3ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16m2))) +void vsoxseg3ei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16m2_m))) +void vsoxseg3ei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16mf2))) +void vsoxseg3ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16mf2_m))) +void vsoxseg3ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16mf4))) +void vsoxseg3ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16mf4_m))) +void vsoxseg3ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16m1))) +void vsoxseg4ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16m1_m))) +void vsoxseg4ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16m2))) +void vsoxseg4ei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16m2_m))) +void vsoxseg4ei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16mf2))) +void vsoxseg4ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16mf2_m))) +void vsoxseg4ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16mf4))) +void vsoxseg4ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16mf4_m))) +void vsoxseg4ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f16m1))) +void vsoxseg5ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f16m1_m))) +void vsoxseg5ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f16mf2))) +void vsoxseg5ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f16mf2_m))) +void vsoxseg5ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f16mf4))) +void vsoxseg5ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f16mf4_m))) +void vsoxseg5ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f16m1))) +void vsoxseg6ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f16m1_m))) +void vsoxseg6ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f16mf2))) +void vsoxseg6ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f16mf2_m))) +void vsoxseg6ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f16mf4))) +void vsoxseg6ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f16mf4_m))) +void vsoxseg6ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f16m1))) +void vsoxseg7ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f16m1_m))) +void vsoxseg7ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f16mf2))) +void vsoxseg7ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f16mf2_m))) +void vsoxseg7ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f16mf4))) +void vsoxseg7ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f16mf4_m))) +void vsoxseg7ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f16m1))) +void vsoxseg8ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f16m1_m))) +void vsoxseg8ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f16mf2))) +void vsoxseg8ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f16mf2_m))) +void vsoxseg8ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f16mf4))) +void vsoxseg8ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f16mf4_m))) +void vsoxseg8ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_f16m1_m))) +void vlseg2e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_f16m2_m))) +void vlseg2e16(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_f16m4_m))) +void vlseg2e16(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_f16mf2_m))) +void vlseg2e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_f16mf4_m))) +void vlseg2e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_f16m1_m))) +void vlseg3e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_f16m2_m))) +void vlseg3e16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_f16mf2_m))) +void vlseg3e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_f16mf4_m))) +void vlseg3e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_f16m1_m))) +void vlseg4e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_f16m2_m))) +void vlseg4e16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_f16mf2_m))) +void vlseg4e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_f16mf4_m))) +void vlseg4e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_f16m1_m))) +void vlseg5e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_f16mf2_m))) +void vlseg5e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_f16mf4_m))) +void vlseg5e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_f16m1_m))) +void vlseg6e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_f16mf2_m))) +void vlseg6e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_f16mf4_m))) +void vlseg6e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_f16m1_m))) +void vlseg7e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_f16mf2_m))) +void vlseg7e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_f16mf4_m))) +void vlseg7e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_f16m1_m))) +void vlseg8e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_f16mf2_m))) +void vlseg8e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_f16mf4_m))) +void vlseg8e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_f16m1_m))) +void vlseg2e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_f16m2_m))) +void vlseg2e16ff(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_f16m4_m))) +void vlseg2e16ff(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_f16mf2_m))) +void vlseg2e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_f16mf4_m))) +void vlseg2e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_f16m1_m))) +void vlseg3e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_f16m2_m))) +void vlseg3e16ff(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_f16mf2_m))) +void vlseg3e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_f16mf4_m))) +void vlseg3e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_f16m1_m))) +void vlseg4e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_f16m2_m))) +void vlseg4e16ff(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_f16mf2_m))) +void vlseg4e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_f16mf4_m))) +void vlseg4e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_f16m1_m))) +void vlseg5e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_f16mf2_m))) +void vlseg5e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_f16mf4_m))) +void vlseg5e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_f16m1_m))) +void vlseg6e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_f16mf2_m))) +void vlseg6e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_f16mf4_m))) +void vlseg6e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_f16m1_m))) +void vlseg7e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_f16mf2_m))) +void vlseg7e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_f16mf4_m))) +void vlseg7e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_f16m1_m))) +void vlseg8e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_f16mf2_m))) +void vlseg8e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_f16mf4_m))) +void vlseg8e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_f16m1_m))) +void vlsseg2e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_f16m2_m))) +void vlsseg2e16(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_f16m4_m))) +void vlsseg2e16(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_f16mf2_m))) +void vlsseg2e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_f16mf4_m))) +void vlsseg2e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_f16m1_m))) +void vlsseg3e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_f16m2_m))) +void vlsseg3e16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_f16mf2_m))) +void vlsseg3e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_f16mf4_m))) +void vlsseg3e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_f16m1_m))) +void vlsseg4e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_f16m2_m))) +void vlsseg4e16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_f16mf2_m))) +void vlsseg4e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_f16mf4_m))) +void vlsseg4e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_f16m1_m))) +void vlsseg5e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_f16mf2_m))) +void vlsseg5e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_f16mf4_m))) +void vlsseg5e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_f16m1_m))) +void vlsseg6e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_f16mf2_m))) +void vlsseg6e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_f16mf4_m))) +void vlsseg6e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_f16m1_m))) +void vlsseg7e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_f16mf2_m))) +void vlsseg7e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_f16mf4_m))) +void vlsseg7e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_f16m1_m))) +void vlsseg8e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_f16mf2_m))) +void vlsseg8e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_f16mf4_m))) +void vlsseg8e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16m1))) +void vluxseg2ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16m1_m))) +void vluxseg2ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16m2))) +void vluxseg2ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16m2_m))) +void vluxseg2ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16m4))) +void vluxseg2ei8(vfloat16m4_t * op0, vfloat16m4_t * op1, const _Float16 * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16m4_m))) +void vluxseg2ei8(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16mf2))) +void vluxseg2ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16mf2_m))) +void vluxseg2ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16mf4))) +void vluxseg2ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16mf4_m))) +void vluxseg2ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16m1))) +void vluxseg3ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16m1_m))) +void vluxseg3ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16m2))) +void vluxseg3ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16m2_m))) +void vluxseg3ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16mf2))) +void vluxseg3ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16mf2_m))) +void vluxseg3ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16mf4))) +void vluxseg3ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16mf4_m))) +void vluxseg3ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16m1))) +void vluxseg4ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16m1_m))) +void vluxseg4ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16m2))) +void vluxseg4ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16m2_m))) +void vluxseg4ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16mf2))) +void vluxseg4ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16mf2_m))) +void vluxseg4ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16mf4))) +void vluxseg4ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16mf4_m))) +void vluxseg4ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f16m1))) +void vluxseg5ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f16m1_m))) +void vluxseg5ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f16mf2))) +void vluxseg5ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f16mf2_m))) +void vluxseg5ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f16mf4))) +void vluxseg5ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f16mf4_m))) +void vluxseg5ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f16m1))) +void vluxseg6ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f16m1_m))) +void vluxseg6ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f16mf2))) +void vluxseg6ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f16mf2_m))) +void vluxseg6ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f16mf4))) +void vluxseg6ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f16mf4_m))) +void vluxseg6ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f16m1))) +void vluxseg7ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f16m1_m))) +void vluxseg7ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f16mf2))) +void vluxseg7ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f16mf2_m))) +void vluxseg7ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f16mf4))) +void vluxseg7ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f16mf4_m))) +void vluxseg7ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f16m1))) +void vluxseg8ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f16m1_m))) +void vluxseg8ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f16mf2))) +void vluxseg8ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f16mf2_m))) +void vluxseg8ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f16mf4))) +void vluxseg8ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f16mf4_m))) +void vluxseg8ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16m1))) +void vluxseg2ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16m1_m))) +void vluxseg2ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16m2))) +void vluxseg2ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16m2_m))) +void vluxseg2ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16m4))) +void vluxseg2ei16(vfloat16m4_t * op0, vfloat16m4_t * op1, const _Float16 * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16m4_m))) +void vluxseg2ei16(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16mf2))) +void vluxseg2ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16mf2_m))) +void vluxseg2ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16mf4))) +void vluxseg2ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16mf4_m))) +void vluxseg2ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16m1))) +void vluxseg3ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16m1_m))) +void vluxseg3ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16m2))) +void vluxseg3ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16m2_m))) +void vluxseg3ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16mf2))) +void vluxseg3ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16mf2_m))) +void vluxseg3ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16mf4))) +void vluxseg3ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16mf4_m))) +void vluxseg3ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16m1))) +void vluxseg4ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16m1_m))) +void vluxseg4ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16m2))) +void vluxseg4ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16m2_m))) +void vluxseg4ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16mf2))) +void vluxseg4ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16mf2_m))) +void vluxseg4ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16mf4))) +void vluxseg4ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16mf4_m))) +void vluxseg4ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f16m1))) +void vluxseg5ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f16m1_m))) +void vluxseg5ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f16mf2))) +void vluxseg5ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f16mf2_m))) +void vluxseg5ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f16mf4))) +void vluxseg5ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f16mf4_m))) +void vluxseg5ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f16m1))) +void vluxseg6ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f16m1_m))) +void vluxseg6ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f16mf2))) +void vluxseg6ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f16mf2_m))) +void vluxseg6ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f16mf4))) +void vluxseg6ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f16mf4_m))) +void vluxseg6ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f16m1))) +void vluxseg7ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f16m1_m))) +void vluxseg7ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f16mf2))) +void vluxseg7ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f16mf2_m))) +void vluxseg7ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f16mf4))) +void vluxseg7ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f16mf4_m))) +void vluxseg7ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f16m1))) +void vluxseg8ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f16m1_m))) +void vluxseg8ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f16mf2))) +void vluxseg8ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f16mf2_m))) +void vluxseg8ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f16mf4))) +void vluxseg8ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f16mf4_m))) +void vluxseg8ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16m1))) +void vluxseg2ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16m1_m))) +void vluxseg2ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16m2))) +void vluxseg2ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16m2_m))) +void vluxseg2ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16m4))) +void vluxseg2ei32(vfloat16m4_t * op0, vfloat16m4_t * op1, const _Float16 * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16m4_m))) +void vluxseg2ei32(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16mf2))) +void vluxseg2ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16mf2_m))) +void vluxseg2ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16mf4))) +void vluxseg2ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16mf4_m))) +void vluxseg2ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16m1))) +void vluxseg3ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16m1_m))) +void vluxseg3ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16m2))) +void vluxseg3ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16m2_m))) +void vluxseg3ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16mf2))) +void vluxseg3ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16mf2_m))) +void vluxseg3ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16mf4))) +void vluxseg3ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16mf4_m))) +void vluxseg3ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16m1))) +void vluxseg4ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16m1_m))) +void vluxseg4ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16m2))) +void vluxseg4ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16m2_m))) +void vluxseg4ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16mf2))) +void vluxseg4ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16mf2_m))) +void vluxseg4ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16mf4))) +void vluxseg4ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16mf4_m))) +void vluxseg4ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f16m1))) +void vluxseg5ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f16m1_m))) +void vluxseg5ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f16mf2))) +void vluxseg5ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f16mf2_m))) +void vluxseg5ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f16mf4))) +void vluxseg5ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f16mf4_m))) +void vluxseg5ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f16m1))) +void vluxseg6ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f16m1_m))) +void vluxseg6ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f16mf2))) +void vluxseg6ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f16mf2_m))) +void vluxseg6ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f16mf4))) +void vluxseg6ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f16mf4_m))) +void vluxseg6ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f16m1))) +void vluxseg7ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f16m1_m))) +void vluxseg7ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f16mf2))) +void vluxseg7ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f16mf2_m))) +void vluxseg7ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f16mf4))) +void vluxseg7ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f16mf4_m))) +void vluxseg7ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f16m1))) +void vluxseg8ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f16m1_m))) +void vluxseg8ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f16mf2))) +void vluxseg8ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f16mf2_m))) +void vluxseg8ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f16mf4))) +void vluxseg8ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f16mf4_m))) +void vluxseg8ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16m1))) +void vluxseg2ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16m1_m))) +void vluxseg2ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16m2))) +void vluxseg2ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16m2_m))) +void vluxseg2ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16mf2))) +void vluxseg2ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16mf2_m))) +void vluxseg2ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16mf4))) +void vluxseg2ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16mf4_m))) +void vluxseg2ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16m1))) +void vluxseg3ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16m1_m))) +void vluxseg3ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16m2))) +void vluxseg3ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16m2_m))) +void vluxseg3ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16mf2))) +void vluxseg3ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16mf2_m))) +void vluxseg3ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16mf4))) +void vluxseg3ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16mf4_m))) +void vluxseg3ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16m1))) +void vluxseg4ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16m1_m))) +void vluxseg4ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16m2))) +void vluxseg4ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16m2_m))) +void vluxseg4ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16mf2))) +void vluxseg4ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16mf2_m))) +void vluxseg4ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16mf4))) +void vluxseg4ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16mf4_m))) +void vluxseg4ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f16m1))) +void vluxseg5ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f16m1_m))) +void vluxseg5ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f16mf2))) +void vluxseg5ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f16mf2_m))) +void vluxseg5ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f16mf4))) +void vluxseg5ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f16mf4_m))) +void vluxseg5ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f16m1))) +void vluxseg6ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f16m1_m))) +void vluxseg6ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f16mf2))) +void vluxseg6ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f16mf2_m))) +void vluxseg6ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f16mf4))) +void vluxseg6ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f16mf4_m))) +void vluxseg6ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f16m1))) +void vluxseg7ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f16m1_m))) +void vluxseg7ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f16mf2))) +void vluxseg7ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f16mf2_m))) +void vluxseg7ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f16mf4))) +void vluxseg7ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f16mf4_m))) +void vluxseg7ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f16m1))) +void vluxseg8ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f16m1_m))) +void vluxseg8ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f16mf2))) +void vluxseg8ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f16mf2_m))) +void vluxseg8ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f16mf4))) +void vluxseg8ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f16mf4_m))) +void vluxseg8ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint64m1_t op18, size_t op19); + +#endif + + +#ifdef __cplusplus +} +#endif // __riscv_vector +#endif // __RISCV_VECTOR_H diff --git a/lib/include/uintrintrin.h b/lib/include/uintrintrin.h index 78aa8779c3..e3839dcebe 100644 --- a/lib/include/uintrintrin.h +++ b/lib/include/uintrintrin.h @@ -20,6 +20,13 @@ #ifdef __x86_64__ +struct __uintr_frame +{ + unsigned long long rip; + unsigned long long rflags; + unsigned long long rsp; +}; + /// Clears the user interrupt flag (UIF). Its effect takes place immediately: a /// user interrupt cannot be delivered on the instruction boundary following /// CLUI. Can be executed only if CR4.UINT = 1, the logical processor is in diff --git a/lib/include/vaesintrin.h b/lib/include/vaesintrin.h index c4d5c3e751..f3c0807bb9 100644 --- a/lib/include/vaesintrin.h +++ b/lib/include/vaesintrin.h @@ -28,13 +28,6 @@ static __inline__ __m256i __DEFAULT_FN_ATTRS (__v4di) __B); } -static __inline__ __m512i __DEFAULT_FN_ATTRS_F - _mm512_aesenc_epi128(__m512i __A, __m512i __B) -{ - return (__m512i) __builtin_ia32_aesenc512((__v8di) __A, - (__v8di) __B); -} - static __inline__ __m256i __DEFAULT_FN_ATTRS _mm256_aesdec_epi128(__m256i __A, __m256i __B) { @@ -42,13 +35,6 @@ static __inline__ __m256i __DEFAULT_FN_ATTRS (__v4di) __B); } -static __inline__ __m512i __DEFAULT_FN_ATTRS_F - _mm512_aesdec_epi128(__m512i __A, __m512i __B) -{ - return (__m512i) __builtin_ia32_aesdec512((__v8di) __A, - (__v8di) __B); -} - static __inline__ __m256i __DEFAULT_FN_ATTRS _mm256_aesenclast_epi128(__m256i __A, __m256i __B) { @@ -56,13 +42,6 @@ static __inline__ __m256i __DEFAULT_FN_ATTRS (__v4di) __B); } -static __inline__ __m512i __DEFAULT_FN_ATTRS_F - _mm512_aesenclast_epi128(__m512i __A, __m512i __B) -{ - return (__m512i) __builtin_ia32_aesenclast512((__v8di) __A, - (__v8di) __B); -} - static __inline__ __m256i __DEFAULT_FN_ATTRS _mm256_aesdeclast_epi128(__m256i __A, __m256i __B) { @@ -70,13 +49,35 @@ static __inline__ __m256i __DEFAULT_FN_ATTRS (__v4di) __B); } +#ifdef __AVX512FINTRIN_H +static __inline__ __m512i __DEFAULT_FN_ATTRS_F + _mm512_aesenc_epi128(__m512i __A, __m512i __B) +{ + return (__m512i) __builtin_ia32_aesenc512((__v8di) __A, + (__v8di) __B); +} + +static __inline__ __m512i __DEFAULT_FN_ATTRS_F + _mm512_aesdec_epi128(__m512i __A, __m512i __B) +{ + return (__m512i) __builtin_ia32_aesdec512((__v8di) __A, + (__v8di) __B); +} + +static __inline__ __m512i __DEFAULT_FN_ATTRS_F + _mm512_aesenclast_epi128(__m512i __A, __m512i __B) +{ + return (__m512i) __builtin_ia32_aesenclast512((__v8di) __A, + (__v8di) __B); +} + static __inline__ __m512i __DEFAULT_FN_ATTRS_F _mm512_aesdeclast_epi128(__m512i __A, __m512i __B) { return (__m512i) __builtin_ia32_aesdeclast512((__v8di) __A, (__v8di) __B); } - +#endif // __AVX512FINTRIN_H #undef __DEFAULT_FN_ATTRS #undef __DEFAULT_FN_ATTRS_F diff --git a/lib/include/vecintrin.h b/lib/include/vecintrin.h index e58c9769e8..ec1dbfd015 100644 --- a/lib/include/vecintrin.h +++ b/lib/include/vecintrin.h @@ -1016,64 +1016,84 @@ vec_scatter_element(__vector double __vec, static inline __ATTRS_o_ai __vector signed char vec_xl(long __offset, const signed char *__ptr) { - return *(const __vector signed char *) - ((const char *)__ptr + __offset); + __vector signed char V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed char)); + return V; } static inline __ATTRS_o_ai __vector unsigned char vec_xl(long __offset, const unsigned char *__ptr) { - return *(const __vector unsigned char *) - ((const char *)__ptr + __offset); + __vector unsigned char V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned char)); + return V; } static inline __ATTRS_o_ai __vector signed short vec_xl(long __offset, const signed short *__ptr) { - return *(const __vector signed short *) - ((const char *)__ptr + __offset); + __vector signed short V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed short)); + return V; } static inline __ATTRS_o_ai __vector unsigned short vec_xl(long __offset, const unsigned short *__ptr) { - return *(const __vector unsigned short *) - ((const char *)__ptr + __offset); + __vector unsigned short V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned short)); + return V; } static inline __ATTRS_o_ai __vector signed int vec_xl(long __offset, const signed int *__ptr) { - return *(const __vector signed int *) - ((const char *)__ptr + __offset); + __vector signed int V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed int)); + return V; } static inline __ATTRS_o_ai __vector unsigned int vec_xl(long __offset, const unsigned int *__ptr) { - return *(const __vector unsigned int *) - ((const char *)__ptr + __offset); + __vector unsigned int V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned int)); + return V; } static inline __ATTRS_o_ai __vector signed long long vec_xl(long __offset, const signed long long *__ptr) { - return *(const __vector signed long long *) - ((const char *)__ptr + __offset); + __vector signed long long V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed long long)); + return V; } static inline __ATTRS_o_ai __vector unsigned long long vec_xl(long __offset, const unsigned long long *__ptr) { - return *(const __vector unsigned long long *) - ((const char *)__ptr + __offset); + __vector unsigned long long V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned long long)); + return V; } #if __ARCH__ >= 12 static inline __ATTRS_o_ai __vector float vec_xl(long __offset, const float *__ptr) { - return *(const __vector float *) - ((const char *)__ptr + __offset); + __vector float V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector float)); + return V; } #endif static inline __ATTRS_o_ai __vector double vec_xl(long __offset, const double *__ptr) { - return *(const __vector double *) - ((const char *)__ptr + __offset); + __vector double V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector double)); + return V; } /*-- vec_xld2 ---------------------------------------------------------------*/ @@ -1081,64 +1101,82 @@ vec_xl(long __offset, const double *__ptr) { // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed char vec_xld2(long __offset, const signed char *__ptr) { - return *(const __vector signed char *) - ((const char *)__ptr + __offset); + __vector signed char V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed char)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned char vec_xld2(long __offset, const unsigned char *__ptr) { - return *(const __vector unsigned char *) - ((const char *)__ptr + __offset); + __vector unsigned char V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned char)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed short vec_xld2(long __offset, const signed short *__ptr) { - return *(const __vector signed short *) - ((const char *)__ptr + __offset); + __vector signed short V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed short)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned short vec_xld2(long __offset, const unsigned short *__ptr) { - return *(const __vector unsigned short *) - ((const char *)__ptr + __offset); + __vector unsigned short V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned short)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed int vec_xld2(long __offset, const signed int *__ptr) { - return *(const __vector signed int *) - ((const char *)__ptr + __offset); + __vector signed int V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed int)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned int vec_xld2(long __offset, const unsigned int *__ptr) { - return *(const __vector unsigned int *) - ((const char *)__ptr + __offset); + __vector unsigned int V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned int)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed long long vec_xld2(long __offset, const signed long long *__ptr) { - return *(const __vector signed long long *) - ((const char *)__ptr + __offset); + __vector signed long long V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed long long)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned long long vec_xld2(long __offset, const unsigned long long *__ptr) { - return *(const __vector unsigned long long *) - ((const char *)__ptr + __offset); + __vector unsigned long long V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned long long)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector double vec_xld2(long __offset, const double *__ptr) { - return *(const __vector double *) - ((const char *)__ptr + __offset); + __vector double V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector double)); + return V; } /*-- vec_xlw4 ---------------------------------------------------------------*/ @@ -1146,99 +1184,128 @@ vec_xld2(long __offset, const double *__ptr) { // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed char vec_xlw4(long __offset, const signed char *__ptr) { - return *(const __vector signed char *) - ((const char *)__ptr + __offset); + __vector signed char V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed char)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned char vec_xlw4(long __offset, const unsigned char *__ptr) { - return *(const __vector unsigned char *) - ((const char *)__ptr + __offset); + __vector unsigned char V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned char)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed short vec_xlw4(long __offset, const signed short *__ptr) { - return *(const __vector signed short *) - ((const char *)__ptr + __offset); + __vector signed short V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed short)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned short vec_xlw4(long __offset, const unsigned short *__ptr) { - return *(const __vector unsigned short *) - ((const char *)__ptr + __offset); + __vector unsigned short V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned short)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed int vec_xlw4(long __offset, const signed int *__ptr) { - return *(const __vector signed int *) - ((const char *)__ptr + __offset); + __vector signed int V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed int)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned int vec_xlw4(long __offset, const unsigned int *__ptr) { - return *(const __vector unsigned int *) - ((const char *)__ptr + __offset); + __vector unsigned int V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned int)); + return V; } /*-- vec_xst ----------------------------------------------------------------*/ static inline __ATTRS_o_ai void vec_xst(__vector signed char __vec, long __offset, signed char *__ptr) { - *(__vector signed char *)((char *)__ptr + __offset) = __vec; + __vector signed char V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed char)); } static inline __ATTRS_o_ai void vec_xst(__vector unsigned char __vec, long __offset, unsigned char *__ptr) { - *(__vector unsigned char *)((char *)__ptr + __offset) = __vec; + __vector unsigned char V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned char)); } static inline __ATTRS_o_ai void vec_xst(__vector signed short __vec, long __offset, signed short *__ptr) { - *(__vector signed short *)((char *)__ptr + __offset) = __vec; + __vector signed short V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed short)); } static inline __ATTRS_o_ai void vec_xst(__vector unsigned short __vec, long __offset, unsigned short *__ptr) { - *(__vector unsigned short *)((char *)__ptr + __offset) = __vec; + __vector unsigned short V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned short)); } static inline __ATTRS_o_ai void vec_xst(__vector signed int __vec, long __offset, signed int *__ptr) { - *(__vector signed int *)((char *)__ptr + __offset) = __vec; + __vector signed int V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, sizeof(__vector signed int)); } static inline __ATTRS_o_ai void vec_xst(__vector unsigned int __vec, long __offset, unsigned int *__ptr) { - *(__vector unsigned int *)((char *)__ptr + __offset) = __vec; + __vector unsigned int V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned int)); } static inline __ATTRS_o_ai void vec_xst(__vector signed long long __vec, long __offset, - signed long long *__ptr) { - *(__vector signed long long *)((char *)__ptr + __offset) = __vec; + signed long long *__ptr) { + __vector signed long long V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed long long)); } static inline __ATTRS_o_ai void vec_xst(__vector unsigned long long __vec, long __offset, - unsigned long long *__ptr) { - *(__vector unsigned long long *)((char *)__ptr + __offset) = __vec; + unsigned long long *__ptr) { + __vector unsigned long long V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned long long)); } #if __ARCH__ >= 12 static inline __ATTRS_o_ai void vec_xst(__vector float __vec, long __offset, float *__ptr) { - *(__vector float *)((char *)__ptr + __offset) = __vec; + __vector float V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, sizeof(__vector float)); } #endif static inline __ATTRS_o_ai void vec_xst(__vector double __vec, long __offset, double *__ptr) { - *(__vector double *)((char *)__ptr + __offset) = __vec; + __vector double V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, sizeof(__vector double)); } /*-- vec_xstd2 --------------------------------------------------------------*/ @@ -1246,57 +1313,73 @@ vec_xst(__vector double __vec, long __offset, double *__ptr) { // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector signed char __vec, long __offset, signed char *__ptr) { - *(__vector signed char *)((char *)__ptr + __offset) = __vec; + __vector signed char V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed char)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector unsigned char __vec, long __offset, unsigned char *__ptr) { - *(__vector unsigned char *)((char *)__ptr + __offset) = __vec; + __vector unsigned char V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned char)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector signed short __vec, long __offset, signed short *__ptr) { - *(__vector signed short *)((char *)__ptr + __offset) = __vec; + __vector signed short V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed short)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector unsigned short __vec, long __offset, unsigned short *__ptr) { - *(__vector unsigned short *)((char *)__ptr + __offset) = __vec; + __vector unsigned short V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned short)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector signed int __vec, long __offset, signed int *__ptr) { - *(__vector signed int *)((char *)__ptr + __offset) = __vec; + __vector signed int V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, sizeof(__vector signed int)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector unsigned int __vec, long __offset, unsigned int *__ptr) { - *(__vector unsigned int *)((char *)__ptr + __offset) = __vec; + __vector unsigned int V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned int)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector signed long long __vec, long __offset, signed long long *__ptr) { - *(__vector signed long long *)((char *)__ptr + __offset) = __vec; + __vector signed long long V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed long long)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector unsigned long long __vec, long __offset, unsigned long long *__ptr) { - *(__vector unsigned long long *)((char *)__ptr + __offset) = __vec; + __vector unsigned long long V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned long long)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector double __vec, long __offset, double *__ptr) { - *(__vector double *)((char *)__ptr + __offset) = __vec; + __vector double V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, sizeof(__vector double)); } /*-- vec_xstw4 --------------------------------------------------------------*/ @@ -1304,37 +1387,48 @@ vec_xstd2(__vector double __vec, long __offset, double *__ptr) { // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstw4(__vector signed char __vec, long __offset, signed char *__ptr) { - *(__vector signed char *)((char *)__ptr + __offset) = __vec; + __vector signed char V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed char)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstw4(__vector unsigned char __vec, long __offset, unsigned char *__ptr) { - *(__vector unsigned char *)((char *)__ptr + __offset) = __vec; + __vector unsigned char V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned char)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstw4(__vector signed short __vec, long __offset, signed short *__ptr) { - *(__vector signed short *)((char *)__ptr + __offset) = __vec; + __vector signed short V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed short)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstw4(__vector unsigned short __vec, long __offset, unsigned short *__ptr) { - *(__vector unsigned short *)((char *)__ptr + __offset) = __vec; + __vector unsigned short V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned short)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstw4(__vector signed int __vec, long __offset, signed int *__ptr) { - *(__vector signed int *)((char *)__ptr + __offset) = __vec; + __vector signed int V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, sizeof(__vector signed int)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstw4(__vector unsigned int __vec, long __offset, unsigned int *__ptr) { - *(__vector unsigned int *)((char *)__ptr + __offset) = __vec; + __vector unsigned int V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned int)); } /*-- vec_load_bndry ---------------------------------------------------------*/ @@ -9259,6 +9353,41 @@ vec_fp_test_data_class(__vector double __a, int __b, int *__c) __VEC_CLASS_FP_ZERO | \ __VEC_CLASS_FP_INFINITY) +/*-- vec_extend_to_fp32_hi --------------------------------------------------*/ + +#if __ARCH__ >= 14 +#define vec_extend_to_fp32_hi(X, W) \ + ((__vector float)__builtin_s390_vclfnhs((X), (W))); +#endif + +/*-- vec_extend_to_fp32_hi --------------------------------------------------*/ + +#if __ARCH__ >= 14 +#define vec_extend_to_fp32_lo(X, W) \ + ((__vector float)__builtin_s390_vclfnls((X), (W))); +#endif + +/*-- vec_round_from_fp32 ----------------------------------------------------*/ + +#if __ARCH__ >= 14 +#define vec_round_from_fp32(X, Y, W) \ + ((__vector unsigned short)__builtin_s390_vcrnfs((X), (Y), (W))); +#endif + +/*-- vec_convert_to_fp16 ----------------------------------------------------*/ + +#if __ARCH__ >= 14 +#define vec_convert_to_fp16(X, W) \ + ((__vector unsigned short)__builtin_s390_vcfn((X), (W))); +#endif + +/*-- vec_convert_from_fp16 --------------------------------------------------*/ + +#if __ARCH__ >= 14 +#define vec_convert_from_fp16(X, W) \ + ((__vector unsigned short)__builtin_s390_vcnf((X), (W))); +#endif + /*-- vec_cp_until_zero ------------------------------------------------------*/ static inline __ATTRS_o_ai __vector signed char diff --git a/lib/include/vpclmulqdqintrin.h b/lib/include/vpclmulqdqintrin.h index 470d832549..44daadb07d 100644 --- a/lib/include/vpclmulqdqintrin.h +++ b/lib/include/vpclmulqdqintrin.h @@ -19,10 +19,12 @@ (__v4di)(__m256i)(B), \ (char)(I)) +#ifdef __AVX512FINTRIN_H #define _mm512_clmulepi64_epi128(A, B, I) \ (__m512i)__builtin_ia32_pclmulqdq512((__v8di)(__m512i)(A), \ (__v8di)(__m512i)(B), \ (char)(I)) +#endif // __AVX512FINTRIN_H #endif /* __VPCLMULQDQINTRIN_H */ diff --git a/lib/include/wasm_simd128.h b/lib/include/wasm_simd128.h index ac88516ac9..712fa03780 100644 --- a/lib/include/wasm_simd128.h +++ b/lib/include/wasm_simd128.h @@ -40,13 +40,18 @@ typedef unsigned char __u8x8 typedef short __i16x4 __attribute__((__vector_size__(8), __aligned__(8))); typedef unsigned short __u16x4 __attribute__((__vector_size__(8), __aligned__(8))); +typedef int __i32x2 __attribute__((__vector_size__(8), __aligned__(8))); +typedef unsigned int __u32x2 + __attribute__((__vector_size__(8), __aligned__(8))); +typedef float __f32x2 __attribute__((__vector_size__(8), __aligned__(8))); #define __DEFAULT_FN_ATTRS \ __attribute__((__always_inline__, __nodebug__, __target__("simd128"), \ __min_vector_width__(128))) -#define __REQUIRE_CONSTANT(e) \ - _Static_assert(__builtin_constant_p(e), "Expected constant") +#define __REQUIRE_CONSTANT(c) \ + __attribute__((__diagnose_if__(!__builtin_constant_p(c), \ + #c " must be constant", "error"))) static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_load(const void *__mem) { // UB-free unaligned access copied from xmmintrin.h @@ -57,102 +62,158 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_load(const void *__mem) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_v8x16_load_splat(const void *__mem) { - struct __wasm_v8x16_load_splat_struct { +wasm_v128_load8_splat(const void *__mem) { + struct __wasm_v128_load8_splat_struct { uint8_t __v; } __attribute__((__packed__, __may_alias__)); - uint8_t __v = ((const struct __wasm_v8x16_load_splat_struct *)__mem)->__v; + uint8_t __v = ((const struct __wasm_v128_load8_splat_struct *)__mem)->__v; return (v128_t)(__u8x16){__v, __v, __v, __v, __v, __v, __v, __v, __v, __v, __v, __v, __v, __v, __v, __v}; } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_v16x8_load_splat(const void *__mem) { - struct __wasm_v16x8_load_splat_struct { +wasm_v128_load16_splat(const void *__mem) { + struct __wasm_v128_load16_splat_struct { uint16_t __v; } __attribute__((__packed__, __may_alias__)); - uint16_t __v = ((const struct __wasm_v16x8_load_splat_struct *)__mem)->__v; + uint16_t __v = ((const struct __wasm_v128_load16_splat_struct *)__mem)->__v; return (v128_t)(__u16x8){__v, __v, __v, __v, __v, __v, __v, __v}; } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_v32x4_load_splat(const void *__mem) { - struct __wasm_v32x4_load_splat_struct { +wasm_v128_load32_splat(const void *__mem) { + struct __wasm_v128_load32_splat_struct { uint32_t __v; } __attribute__((__packed__, __may_alias__)); - uint32_t __v = ((const struct __wasm_v32x4_load_splat_struct *)__mem)->__v; + uint32_t __v = ((const struct __wasm_v128_load32_splat_struct *)__mem)->__v; return (v128_t)(__u32x4){__v, __v, __v, __v}; } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_v64x2_load_splat(const void *__mem) { - struct __wasm_v64x2_load_splat_struct { +wasm_v128_load64_splat(const void *__mem) { + struct __wasm_v128_load64_splat_struct { uint64_t __v; } __attribute__((__packed__, __may_alias__)); - uint64_t __v = ((const struct __wasm_v64x2_load_splat_struct *)__mem)->__v; + uint64_t __v = ((const struct __wasm_v128_load64_splat_struct *)__mem)->__v; return (v128_t)(__u64x2){__v, __v}; } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_load_8x8(const void *__mem) { - typedef int8_t __i8x8 __attribute__((__vector_size__(8), __aligned__(8))); - struct __wasm_i16x8_load_8x8_struct { +wasm_i16x8_load8x8(const void *__mem) { + struct __wasm_i16x8_load8x8_struct { __i8x8 __v; } __attribute__((__packed__, __may_alias__)); - __i8x8 __v = ((const struct __wasm_i16x8_load_8x8_struct *)__mem)->__v; + __i8x8 __v = ((const struct __wasm_i16x8_load8x8_struct *)__mem)->__v; return (v128_t) __builtin_convertvector(__v, __i16x8); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u16x8_load_8x8(const void *__mem) { - typedef uint8_t __u8x8 __attribute__((__vector_size__(8), __aligned__(8))); - struct __wasm_u16x8_load_8x8_struct { +wasm_u16x8_load8x8(const void *__mem) { + struct __wasm_u16x8_load8x8_struct { __u8x8 __v; } __attribute__((__packed__, __may_alias__)); - __u8x8 __v = ((const struct __wasm_u16x8_load_8x8_struct *)__mem)->__v; + __u8x8 __v = ((const struct __wasm_u16x8_load8x8_struct *)__mem)->__v; return (v128_t) __builtin_convertvector(__v, __u16x8); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i32x4_load_16x4(const void *__mem) { - typedef int16_t __i16x4 __attribute__((__vector_size__(8), __aligned__(8))); - struct __wasm_i32x4_load_16x4_struct { +wasm_i32x4_load16x4(const void *__mem) { + struct __wasm_i32x4_load16x4_struct { __i16x4 __v; } __attribute__((__packed__, __may_alias__)); - __i16x4 __v = ((const struct __wasm_i32x4_load_16x4_struct *)__mem)->__v; + __i16x4 __v = ((const struct __wasm_i32x4_load16x4_struct *)__mem)->__v; return (v128_t) __builtin_convertvector(__v, __i32x4); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u32x4_load_16x4(const void *__mem) { - typedef uint16_t __u16x4 __attribute__((__vector_size__(8), __aligned__(8))); - struct __wasm_u32x4_load_16x4_struct { +wasm_u32x4_load16x4(const void *__mem) { + struct __wasm_u32x4_load16x4_struct { __u16x4 __v; } __attribute__((__packed__, __may_alias__)); - __u16x4 __v = ((const struct __wasm_u32x4_load_16x4_struct *)__mem)->__v; + __u16x4 __v = ((const struct __wasm_u32x4_load16x4_struct *)__mem)->__v; return (v128_t) __builtin_convertvector(__v, __u32x4); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i64x2_load_32x2(const void *__mem) { - typedef int32_t __i32x2 __attribute__((__vector_size__(8), __aligned__(8))); - struct __wasm_i64x2_load_32x2_struct { +wasm_i64x2_load32x2(const void *__mem) { + struct __wasm_i64x2_load32x2_struct { __i32x2 __v; } __attribute__((__packed__, __may_alias__)); - __i32x2 __v = ((const struct __wasm_i64x2_load_32x2_struct *)__mem)->__v; + __i32x2 __v = ((const struct __wasm_i64x2_load32x2_struct *)__mem)->__v; return (v128_t) __builtin_convertvector(__v, __i64x2); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u64x2_load_32x2(const void *__mem) { - typedef uint32_t __u32x2 __attribute__((__vector_size__(8), __aligned__(8))); - struct __wasm_u64x2_load_32x2_struct { +wasm_u64x2_load32x2(const void *__mem) { + struct __wasm_u64x2_load32x2_struct { __u32x2 __v; } __attribute__((__packed__, __may_alias__)); - __u32x2 __v = ((const struct __wasm_u64x2_load_32x2_struct *)__mem)->__v; + __u32x2 __v = ((const struct __wasm_u64x2_load32x2_struct *)__mem)->__v; return (v128_t) __builtin_convertvector(__v, __u64x2); } +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_v128_load32_zero(const void *__mem) { + struct __wasm_v128_load32_zero_struct { + int32_t __v; + } __attribute__((__packed__, __may_alias__)); + int32_t __v = ((const struct __wasm_v128_load32_zero_struct *)__mem)->__v; + return (v128_t)(__i32x4){__v, 0, 0, 0}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_v128_load64_zero(const void *__mem) { + struct __wasm_v128_load64_zero_struct { + int64_t __v; + } __attribute__((__packed__, __may_alias__)); + int64_t __v = ((const struct __wasm_v128_load64_zero_struct *)__mem)->__v; + return (v128_t)(__i64x2){__v, 0}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_load8_lane( + const void *__mem, v128_t __vec, int __i) __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_load8_lane_struct { + int8_t __v; + } __attribute__((__packed__, __may_alias__)); + int8_t __v = ((const struct __wasm_v128_load8_lane_struct *)__mem)->__v; + __i8x16 __ret = (__i8x16)__vec; + __ret[__i] = __v; + return (v128_t)__ret; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_load16_lane( + const void *__mem, v128_t __vec, int __i) __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_load16_lane_struct { + int16_t __v; + } __attribute__((__packed__, __may_alias__)); + int16_t __v = ((const struct __wasm_v128_load16_lane_struct *)__mem)->__v; + __i16x8 __ret = (__i16x8)__vec; + __ret[__i] = __v; + return (v128_t)__ret; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_load32_lane( + const void *__mem, v128_t __vec, int __i) __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_load32_lane_struct { + int32_t __v; + } __attribute__((__packed__, __may_alias__)); + int32_t __v = ((const struct __wasm_v128_load32_lane_struct *)__mem)->__v; + __i32x4 __ret = (__i32x4)__vec; + __ret[__i] = __v; + return (v128_t)__ret; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_load64_lane( + const void *__mem, v128_t __vec, int __i) __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_load64_lane_struct { + int64_t __v; + } __attribute__((__packed__, __may_alias__)); + int64_t __v = ((const struct __wasm_v128_load64_lane_struct *)__mem)->__v; + __i64x2 __ret = (__i64x2)__vec; + __ret[__i] = __v; + return (v128_t)__ret; +} + static __inline__ void __DEFAULT_FN_ATTRS wasm_v128_store(void *__mem, v128_t __a) { // UB-free unaligned access copied from xmmintrin.h @@ -162,6 +223,49 @@ static __inline__ void __DEFAULT_FN_ATTRS wasm_v128_store(void *__mem, ((struct __wasm_v128_store_struct *)__mem)->__v = __a; } +static __inline__ void __DEFAULT_FN_ATTRS wasm_v128_store8_lane(void *__mem, + v128_t __vec, + int __i) + __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_store8_lane_struct { + int8_t __v; + } __attribute__((__packed__, __may_alias__)); + ((struct __wasm_v128_store8_lane_struct *)__mem)->__v = ((__i8x16)__vec)[__i]; +} + +static __inline__ void __DEFAULT_FN_ATTRS wasm_v128_store16_lane(void *__mem, + v128_t __vec, + int __i) + __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_store16_lane_struct { + int16_t __v; + } __attribute__((__packed__, __may_alias__)); + ((struct __wasm_v128_store16_lane_struct *)__mem)->__v = + ((__i16x8)__vec)[__i]; +} + +static __inline__ void __DEFAULT_FN_ATTRS wasm_v128_store32_lane(void *__mem, + v128_t __vec, + int __i) + __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_store32_lane_struct { + int32_t __v; + } __attribute__((__packed__, __may_alias__)); + ((struct __wasm_v128_store32_lane_struct *)__mem)->__v = + ((__i32x4)__vec)[__i]; +} + +static __inline__ void __DEFAULT_FN_ATTRS wasm_v128_store64_lane(void *__mem, + v128_t __vec, + int __i) + __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_store64_lane_struct { + int64_t __v; + } __attribute__((__packed__, __may_alias__)); + ((struct __wasm_v128_store64_lane_struct *)__mem)->__v = + ((__i64x2)__vec)[__i]; +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_make(int8_t __c0, int8_t __c1, int8_t __c2, int8_t __c3, int8_t __c4, int8_t __c5, int8_t __c6, int8_t __c7, int8_t __c8, int8_t __c9, @@ -185,6 +289,11 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_make(int32_t __c0, return (v128_t)(__i32x4){__c0, __c1, __c2, __c3}; } +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_make(int64_t __c0, + int64_t __c1) { + return (v128_t)(__i64x2){__c0, __c1}; +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_make(float __c0, float __c1, float __c2, @@ -192,150 +301,221 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_make(float __c0, return (v128_t)(__f32x4){__c0, __c1, __c2, __c3}; } -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_make(int64_t __c0, - int64_t __c1) { - return (v128_t)(__i64x2){__c0, __c1}; -} - static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_make(double __c0, double __c1) { return (v128_t)(__f64x2){__c0, __c1}; } -#define wasm_i8x16_const(__c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7, __c8, \ - __c9, __c10, __c11, __c12, __c13, __c14, __c15) \ - __extension__({ \ - __REQUIRE_CONSTANT(__c0); \ - __REQUIRE_CONSTANT(__c1); \ - __REQUIRE_CONSTANT(__c2); \ - __REQUIRE_CONSTANT(__c3); \ - __REQUIRE_CONSTANT(__c4); \ - __REQUIRE_CONSTANT(__c5); \ - __REQUIRE_CONSTANT(__c6); \ - __REQUIRE_CONSTANT(__c7); \ - __REQUIRE_CONSTANT(__c8); \ - __REQUIRE_CONSTANT(__c9); \ - __REQUIRE_CONSTANT(__c10); \ - __REQUIRE_CONSTANT(__c11); \ - __REQUIRE_CONSTANT(__c12); \ - __REQUIRE_CONSTANT(__c13); \ - __REQUIRE_CONSTANT(__c14); \ - __REQUIRE_CONSTANT(__c15); \ - (v128_t)(__i8x16){__c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7, \ - __c8, __c9, __c10, __c11, __c12, __c13, __c14, __c15}; \ - }) +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i8x16_const(int8_t __c0, int8_t __c1, int8_t __c2, int8_t __c3, + int8_t __c4, int8_t __c5, int8_t __c6, int8_t __c7, + int8_t __c8, int8_t __c9, int8_t __c10, int8_t __c11, + int8_t __c12, int8_t __c13, int8_t __c14, int8_t __c15) + __REQUIRE_CONSTANT(__c0) __REQUIRE_CONSTANT(__c1) __REQUIRE_CONSTANT(__c2) + __REQUIRE_CONSTANT(__c3) __REQUIRE_CONSTANT(__c4) + __REQUIRE_CONSTANT(__c5) __REQUIRE_CONSTANT(__c6) + __REQUIRE_CONSTANT(__c7) __REQUIRE_CONSTANT(__c8) + __REQUIRE_CONSTANT(__c9) __REQUIRE_CONSTANT(__c10) + __REQUIRE_CONSTANT(__c11) __REQUIRE_CONSTANT(__c12) + __REQUIRE_CONSTANT(__c13) __REQUIRE_CONSTANT(__c14) + __REQUIRE_CONSTANT(__c15) { + return (v128_t)(__i8x16){__c0, __c1, __c2, __c3, __c4, __c5, + __c6, __c7, __c8, __c9, __c10, __c11, + __c12, __c13, __c14, __c15}; +} -#define wasm_i16x8_const(__c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7) \ - __extension__({ \ - __REQUIRE_CONSTANT(__c0); \ - __REQUIRE_CONSTANT(__c1); \ - __REQUIRE_CONSTANT(__c2); \ - __REQUIRE_CONSTANT(__c3); \ - __REQUIRE_CONSTANT(__c4); \ - __REQUIRE_CONSTANT(__c5); \ - __REQUIRE_CONSTANT(__c6); \ - __REQUIRE_CONSTANT(__c7); \ - (v128_t)(__i16x8){__c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7}; \ - }) +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i16x8_const(int16_t __c0, int16_t __c1, int16_t __c2, int16_t __c3, + int16_t __c4, int16_t __c5, int16_t __c6, int16_t __c7) + __REQUIRE_CONSTANT(__c0) __REQUIRE_CONSTANT(__c1) __REQUIRE_CONSTANT(__c2) + __REQUIRE_CONSTANT(__c3) __REQUIRE_CONSTANT(__c4) + __REQUIRE_CONSTANT(__c5) __REQUIRE_CONSTANT(__c6) + __REQUIRE_CONSTANT(__c7) { + return (v128_t)(__i16x8){__c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7}; +} -#define wasm_i32x4_const(__c0, __c1, __c2, __c3) \ - __extension__({ \ - __REQUIRE_CONSTANT(__c0); \ - __REQUIRE_CONSTANT(__c1); \ - __REQUIRE_CONSTANT(__c2); \ - __REQUIRE_CONSTANT(__c3); \ - (v128_t)(__i32x4){__c0, __c1, __c2, __c3}; \ - }) +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i32x4_const(int32_t __c0, int32_t __c1, int32_t __c2, int32_t __c3) + __REQUIRE_CONSTANT(__c0) __REQUIRE_CONSTANT(__c1) __REQUIRE_CONSTANT(__c2) + __REQUIRE_CONSTANT(__c3) { + return (v128_t)(__i32x4){__c0, __c1, __c2, __c3}; +} -#define wasm_f32x4_const(__c0, __c1, __c2, __c3) \ - __extension__({ \ - __REQUIRE_CONSTANT(__c0); \ - __REQUIRE_CONSTANT(__c1); \ - __REQUIRE_CONSTANT(__c2); \ - __REQUIRE_CONSTANT(__c3); \ - (v128_t)(__f32x4){__c0, __c1, __c2, __c3}; \ - }) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_const(int64_t __c0, + int64_t __c1) + __REQUIRE_CONSTANT(__c0) __REQUIRE_CONSTANT(__c1) { + return (v128_t)(__i64x2){__c0, __c1}; +} -#define wasm_i64x2_const(__c0, __c1) \ - __extension__({ \ - __REQUIRE_CONSTANT(__c0); \ - __REQUIRE_CONSTANT(__c1); \ - (v128_t)(__i64x2){__c0, __c1}; \ - }) +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_f32x4_const(float __c0, float __c1, float __c2, float __c3) + __REQUIRE_CONSTANT(__c0) __REQUIRE_CONSTANT(__c1) __REQUIRE_CONSTANT(__c2) + __REQUIRE_CONSTANT(__c3) { + return (v128_t)(__f32x4){__c0, __c1, __c2, __c3}; +} -#define wasm_f64x2_const(__c0, __c1) \ - __extension__({ \ - __REQUIRE_CONSTANT(__c0); \ - __REQUIRE_CONSTANT(__c1); \ - (v128_t)(__f64x2){__c0, __c1}; \ - }) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_const(double __c0, + double __c1) + __REQUIRE_CONSTANT(__c0) __REQUIRE_CONSTANT(__c1) { + return (v128_t)(__f64x2){__c0, __c1}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_const_splat(int8_t __c) + __REQUIRE_CONSTANT(__c) { + return (v128_t)(__i8x16){__c, __c, __c, __c, __c, __c, __c, __c, + __c, __c, __c, __c, __c, __c, __c, __c}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_const_splat(int16_t __c) + __REQUIRE_CONSTANT(__c) { + return (v128_t)(__i16x8){__c, __c, __c, __c, __c, __c, __c, __c}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_const_splat(int32_t __c) + __REQUIRE_CONSTANT(__c) { + return (v128_t)(__i32x4){__c, __c, __c, __c}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_const_splat(int64_t __c) + __REQUIRE_CONSTANT(__c) { + return (v128_t)(__i64x2){__c, __c}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_const_splat(float __c) + __REQUIRE_CONSTANT(__c) { + return (v128_t)(__f32x4){__c, __c, __c, __c}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_const_splat(double __c) + __REQUIRE_CONSTANT(__c) { + return (v128_t)(__f64x2){__c, __c}; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_splat(int8_t __a) { return (v128_t)(__i8x16){__a, __a, __a, __a, __a, __a, __a, __a, __a, __a, __a, __a, __a, __a, __a, __a}; } -#define wasm_i8x16_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_s_i8x16((__i8x16)(__a), __i)) +static __inline__ int8_t __DEFAULT_FN_ATTRS wasm_i8x16_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__i8x16)__a)[__i]; +} -#define wasm_u8x16_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_u_i8x16((__u8x16)(__a), __i)) +static __inline__ uint8_t __DEFAULT_FN_ATTRS wasm_u8x16_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__u8x16)__a)[__i]; +} -#define wasm_i8x16_replace_lane(__a, __i, __b) \ - ((v128_t)__builtin_wasm_replace_lane_i8x16((__i8x16)(__a), __i, __b)) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_replace_lane(v128_t __a, + int __i, + int8_t __b) + __REQUIRE_CONSTANT(__i) { + __i8x16 __v = (__i8x16)__a; + __v[__i] = __b; + return (v128_t)__v; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_splat(int16_t __a) { return (v128_t)(__i16x8){__a, __a, __a, __a, __a, __a, __a, __a}; } -#define wasm_i16x8_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_s_i16x8((__i16x8)(__a), __i)) +static __inline__ int16_t __DEFAULT_FN_ATTRS wasm_i16x8_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__i16x8)__a)[__i]; +} -#define wasm_u16x8_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_u_i16x8((__u16x8)(__a), __i)) +static __inline__ uint16_t __DEFAULT_FN_ATTRS +wasm_u16x8_extract_lane(v128_t __a, int __i) __REQUIRE_CONSTANT(__i) { + return ((__u16x8)__a)[__i]; +} -#define wasm_i16x8_replace_lane(__a, __i, __b) \ - ((v128_t)__builtin_wasm_replace_lane_i16x8((__i16x8)(__a), __i, __b)) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_replace_lane(v128_t __a, + int __i, + int16_t __b) + __REQUIRE_CONSTANT(__i) { + __i16x8 __v = (__i16x8)__a; + __v[__i] = __b; + return (v128_t)__v; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_splat(int32_t __a) { return (v128_t)(__i32x4){__a, __a, __a, __a}; } -#define wasm_i32x4_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_i32x4((__i32x4)(__a), __i)) +static __inline__ int32_t __DEFAULT_FN_ATTRS wasm_i32x4_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__i32x4)__a)[__i]; +} -#define wasm_i32x4_replace_lane(__a, __i, __b) \ - ((v128_t)__builtin_wasm_replace_lane_i32x4((__i32x4)(__a), __i, __b)) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_replace_lane(v128_t __a, + int __i, + int32_t __b) + __REQUIRE_CONSTANT(__i) { + __i32x4 __v = (__i32x4)__a; + __v[__i] = __b; + return (v128_t)__v; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_splat(int64_t __a) { return (v128_t)(__i64x2){__a, __a}; } -#define wasm_i64x2_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_i64x2((__i64x2)(__a), __i)) +static __inline__ int64_t __DEFAULT_FN_ATTRS wasm_i64x2_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__i64x2)__a)[__i]; +} -#define wasm_i64x2_replace_lane(__a, __i, __b) \ - ((v128_t)__builtin_wasm_replace_lane_i64x2((__i64x2)(__a), __i, __b)) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_replace_lane(v128_t __a, + int __i, + int64_t __b) + __REQUIRE_CONSTANT(__i) { + __i64x2 __v = (__i64x2)__a; + __v[__i] = __b; + return (v128_t)__v; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_splat(float __a) { return (v128_t)(__f32x4){__a, __a, __a, __a}; } -#define wasm_f32x4_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_f32x4((__f32x4)(__a), __i)) +static __inline__ float __DEFAULT_FN_ATTRS wasm_f32x4_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__f32x4)__a)[__i]; +} -#define wasm_f32x4_replace_lane(__a, __i, __b) \ - ((v128_t)__builtin_wasm_replace_lane_f32x4((__f32x4)(__a), __i, __b)) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_replace_lane(v128_t __a, + int __i, + float __b) + __REQUIRE_CONSTANT(__i) { + __f32x4 __v = (__f32x4)__a; + __v[__i] = __b; + return (v128_t)__v; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_splat(double __a) { return (v128_t)(__f64x2){__a, __a}; } -#define wasm_f64x2_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_f64x2((__f64x2)(__a), __i)) +static __inline__ double __DEFAULT_FN_ATTRS wasm_f64x2_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__f64x2)__a)[__i]; +} -#define wasm_f64x2_replace_lane(__a, __i, __b) \ - ((v128_t)__builtin_wasm_replace_lane_f64x2((__f64x2)(__a), __i, __b)) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_replace_lane(v128_t __a, + int __i, + double __b) + __REQUIRE_CONSTANT(__i) { + __f64x2 __v = (__f64x2)__a; + __v[__i] = __b; + return (v128_t)__v; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_eq(v128_t __a, v128_t __b) { @@ -487,6 +667,36 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u32x4_ge(v128_t __a, return (v128_t)((__u32x4)__a >= (__u32x4)__b); } +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_eq(v128_t __a, + v128_t __b) { + return (v128_t)((__i64x2)__a == (__i64x2)__b); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_ne(v128_t __a, + v128_t __b) { + return (v128_t)((__i64x2)__a != (__i64x2)__b); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_lt(v128_t __a, + v128_t __b) { + return (v128_t)((__i64x2)__a < (__i64x2)__b); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_gt(v128_t __a, + v128_t __b) { + return (v128_t)((__i64x2)__a > (__i64x2)__b); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_le(v128_t __a, + v128_t __b) { + return (v128_t)((__i64x2)__a <= (__i64x2)__b); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_ge(v128_t __a, + v128_t __b) { + return (v128_t)((__i64x2)__a >= (__i64x2)__b); +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_eq(v128_t __a, v128_t __b) { return (v128_t)((__f32x4)__a == (__f32x4)__b); @@ -571,6 +781,10 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_andnot(v128_t __a, return __a & ~__b; } +static __inline__ bool __DEFAULT_FN_ATTRS wasm_v128_any_true(v128_t __a) { + return __builtin_wasm_any_true_v128((__i8x16)__a); +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_bitselect(v128_t __a, v128_t __b, v128_t __mask) { @@ -586,14 +800,18 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_neg(v128_t __a) { return (v128_t)(-(__u8x16)__a); } -static __inline__ bool __DEFAULT_FN_ATTRS wasm_i8x16_any_true(v128_t __a) { - return __builtin_wasm_any_true_i8x16((__i8x16)__a); -} - static __inline__ bool __DEFAULT_FN_ATTRS wasm_i8x16_all_true(v128_t __a) { return __builtin_wasm_all_true_i8x16((__i8x16)__a); } +static __inline__ int32_t __DEFAULT_FN_ATTRS wasm_i8x16_bitmask(v128_t __a) { + return __builtin_wasm_bitmask_i8x16((__i8x16)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_popcnt(v128_t __a) { + return (v128_t)__builtin_wasm_popcnt_i8x16((__i8x16)__a); +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_shl(v128_t __a, int32_t __b) { return (v128_t)((__i8x16)__a << __b); @@ -614,16 +832,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_add(v128_t __a, return (v128_t)((__u8x16)__a + (__u8x16)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i8x16_add_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_add_saturate_s_i8x16((__i8x16)__a, - (__i8x16)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_add_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_add_sat_s_i8x16((__i8x16)__a, (__i8x16)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u8x16_add_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_add_saturate_u_i8x16((__u8x16)__a, - (__u8x16)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_add_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_add_sat_u_i8x16((__u8x16)__a, (__u8x16)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_sub(v128_t __a, @@ -631,16 +847,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_sub(v128_t __a, return (v128_t)((__u8x16)__a - (__u8x16)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i8x16_sub_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_sub_saturate_s_i8x16((__i8x16)__a, - (__i8x16)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_sub_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_sub_sat_s_i8x16((__i8x16)__a, (__i8x16)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u8x16_sub_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_sub_saturate_u_i8x16((__u8x16)__a, - (__u8x16)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_sub_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_sub_sat_u_i8x16((__u8x16)__a, (__u8x16)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_min(v128_t __a, @@ -676,14 +890,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_neg(v128_t __a) { return (v128_t)(-(__u16x8)__a); } -static __inline__ bool __DEFAULT_FN_ATTRS wasm_i16x8_any_true(v128_t __a) { - return __builtin_wasm_any_true_i16x8((__i16x8)__a); -} - static __inline__ bool __DEFAULT_FN_ATTRS wasm_i16x8_all_true(v128_t __a) { return __builtin_wasm_all_true_i16x8((__i16x8)__a); } +static __inline__ int32_t __DEFAULT_FN_ATTRS wasm_i16x8_bitmask(v128_t __a) { + return __builtin_wasm_bitmask_i16x8((__i16x8)__a); +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_shl(v128_t __a, int32_t __b) { return (v128_t)((__i16x8)__a << __b); @@ -704,16 +918,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_add(v128_t __a, return (v128_t)((__u16x8)__a + (__u16x8)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_add_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_add_saturate_s_i16x8((__i16x8)__a, - (__i16x8)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_add_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_add_sat_s_i16x8((__i16x8)__a, (__i16x8)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u16x8_add_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_add_saturate_u_i16x8((__u16x8)__a, - (__u16x8)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_add_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_add_sat_u_i16x8((__u16x8)__a, (__u16x8)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_sub(v128_t __a, @@ -721,16 +933,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_sub(v128_t __a, return (v128_t)((__i16x8)__a - (__i16x8)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_sub_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_sub_saturate_s_i16x8((__i16x8)__a, - (__i16x8)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_sub_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_sub_sat_s_i16x8((__i16x8)__a, (__i16x8)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u16x8_sub_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_sub_saturate_u_i16x8((__u16x8)__a, - (__u16x8)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_sub_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_sub_sat_u_i16x8((__u16x8)__a, (__u16x8)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_mul(v128_t __a, @@ -771,14 +981,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_neg(v128_t __a) { return (v128_t)(-(__u32x4)__a); } -static __inline__ bool __DEFAULT_FN_ATTRS wasm_i32x4_any_true(v128_t __a) { - return __builtin_wasm_any_true_i32x4((__i32x4)__a); -} - static __inline__ bool __DEFAULT_FN_ATTRS wasm_i32x4_all_true(v128_t __a) { return __builtin_wasm_all_true_i32x4((__i32x4)__a); } +static __inline__ int32_t __DEFAULT_FN_ATTRS wasm_i32x4_bitmask(v128_t __a) { + return __builtin_wasm_bitmask_i32x4((__i32x4)__a); +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_shl(v128_t __a, int32_t __b) { return (v128_t)((__i32x4)__a << __b); @@ -829,21 +1039,26 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u32x4_max(v128_t __a, return (v128_t)__builtin_wasm_max_u_i32x4((__u32x4)__a, (__u32x4)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_neg(v128_t __a) { - return (v128_t)(-(__u64x2)__a); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_dot_i16x8(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_dot_s_i32x4_i16x8((__i16x8)__a, (__i16x8)__b); } -#ifdef __wasm_unimplemented_simd128__ +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_abs(v128_t __a) { + return (v128_t)__builtin_wasm_abs_i64x2((__i64x2)__a); +} -static __inline__ bool __DEFAULT_FN_ATTRS wasm_i64x2_any_true(v128_t __a) { - return __builtin_wasm_any_true_i64x2((__i64x2)__a); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_neg(v128_t __a) { + return (v128_t)(-(__u64x2)__a); } static __inline__ bool __DEFAULT_FN_ATTRS wasm_i64x2_all_true(v128_t __a) { return __builtin_wasm_all_true_i64x2((__i64x2)__a); } -#endif // __wasm_unimplemented_simd128__ +static __inline__ int32_t __DEFAULT_FN_ATTRS wasm_i64x2_bitmask(v128_t __a) { + return __builtin_wasm_bitmask_i64x2((__i64x2)__a); +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_shl(v128_t __a, int32_t __b) { @@ -887,23 +1102,21 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_sqrt(v128_t __a) { return (v128_t)__builtin_wasm_sqrt_f32x4((__f32x4)__a); } -#ifdef __wasm_unimplemented_simd128__ - -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_qfma(v128_t __a, - v128_t __b, - v128_t __c) { - return (v128_t)__builtin_wasm_qfma_f32x4((__f32x4)__a, (__f32x4)__b, - (__f32x4)__c); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_ceil(v128_t __a) { + return (v128_t)__builtin_wasm_ceil_f32x4((__f32x4)__a); } -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_qfms(v128_t __a, - v128_t __b, - v128_t __c) { - return (v128_t)__builtin_wasm_qfms_f32x4((__f32x4)__a, (__f32x4)__b, - (__f32x4)__c); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_floor(v128_t __a) { + return (v128_t)__builtin_wasm_floor_f32x4((__f32x4)__a); } -#endif // __wasm_unimplemented_simd128__ +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_trunc(v128_t __a) { + return (v128_t)__builtin_wasm_trunc_f32x4((__f32x4)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_nearest(v128_t __a) { + return (v128_t)__builtin_wasm_nearest_f32x4((__f32x4)__a); +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_add(v128_t __a, v128_t __b) { @@ -937,12 +1150,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_max(v128_t __a, static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_pmin(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_pmin_f32x4((__f32x4)__a, (__f32x4)__b); + __i32x4 __mask = (__i32x4)((__f32x4)__b < (__f32x4)__a); + return (v128_t)((((__i32x4)__b) & __mask) | (((__i32x4)__a) & ~__mask)); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_pmax(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_pmax_f32x4((__f32x4)__a, (__f32x4)__b); + __i32x4 __mask = (__i32x4)((__f32x4)__a < (__f32x4)__b); + return (v128_t)((((__i32x4)__b) & __mask) | (((__i32x4)__a) & ~__mask)); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_abs(v128_t __a) { @@ -957,23 +1172,21 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_sqrt(v128_t __a) { return (v128_t)__builtin_wasm_sqrt_f64x2((__f64x2)__a); } -#ifdef __wasm_unimplemented_simd128__ - -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_qfma(v128_t __a, - v128_t __b, - v128_t __c) { - return (v128_t)__builtin_wasm_qfma_f64x2((__f64x2)__a, (__f64x2)__b, - (__f64x2)__c); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_ceil(v128_t __a) { + return (v128_t)__builtin_wasm_ceil_f64x2((__f64x2)__a); } -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_qfms(v128_t __a, - v128_t __b, - v128_t __c) { - return (v128_t)__builtin_wasm_qfms_f64x2((__f64x2)__a, (__f64x2)__b, - (__f64x2)__c); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_floor(v128_t __a) { + return (v128_t)__builtin_wasm_floor_f64x2((__f64x2)__a); } -#endif // __wasm_unimplemented_simd128__ +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_trunc(v128_t __a) { + return (v128_t)__builtin_wasm_trunc_f64x2((__f64x2)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_nearest(v128_t __a) { + return (v128_t)__builtin_wasm_nearest_f64x2((__f64x2)__a); +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_add(v128_t __a, v128_t __b) { @@ -1007,21 +1220,23 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_max(v128_t __a, static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_pmin(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_pmin_f64x2((__f64x2)__a, (__f64x2)__b); + __i64x2 __mask = (__i64x2)((__f64x2)__b < (__f64x2)__a); + return (v128_t)((((__i64x2)__b) & __mask) | (((__i64x2)__a) & ~__mask)); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_pmax(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_pmax_f64x2((__f64x2)__a, (__f64x2)__b); + __i64x2 __mask = (__i64x2)((__f64x2)__a < (__f64x2)__b); + return (v128_t)((((__i64x2)__b) & __mask) | (((__i64x2)__a) & ~__mask)); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i32x4_trunc_saturate_f32x4(v128_t __a) { +wasm_i32x4_trunc_sat_f32x4(v128_t __a) { return (v128_t)__builtin_wasm_trunc_saturate_s_i32x4_f32x4((__f32x4)__a); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u32x4_trunc_saturate_f32x4(v128_t __a) { +wasm_u32x4_trunc_sat_f32x4(v128_t __a) { return (v128_t)__builtin_wasm_trunc_saturate_u_i32x4_f32x4((__f32x4)__a); } @@ -1035,38 +1250,71 @@ wasm_f32x4_convert_u32x4(v128_t __a) { return (v128_t) __builtin_convertvector((__u32x4)__a, __f32x4); } -#define wasm_v8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_f64x2_convert_low_i32x4(v128_t __a) { + return (v128_t) __builtin_convertvector((__i32x2){__a[0], __a[1]}, __f64x2); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_f64x2_convert_low_u32x4(v128_t __a) { + return (v128_t) __builtin_convertvector((__u32x2){__a[0], __a[1]}, __f64x2); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i32x4_trunc_sat_f64x2_zero(v128_t __a) { + return (v128_t)__builtin_wasm_trunc_sat_zero_s_f64x2_i32x4((__f64x2)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u32x4_trunc_sat_f64x2_zero(v128_t __a) { + return (v128_t)__builtin_wasm_trunc_sat_zero_u_f64x2_i32x4((__f64x2)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_f32x4_demote_f64x2_zero(v128_t __a) { + return (v128_t) __builtin_convertvector( + __builtin_shufflevector((__f64x2)__a, (__f64x2){0, 0}, 0, 1, 2, 3), + __f32x4); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_f64x2_promote_low_f32x4(v128_t __a) { + return (v128_t) __builtin_convertvector( + (__f32x2){((__f32x4)__a)[0], ((__f32x4)__a)[1]}, __f64x2); +} + +#define wasm_i8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ __c7, __c8, __c9, __c10, __c11, __c12, __c13, \ __c14, __c15) \ - ((v128_t)__builtin_wasm_shuffle_v8x16( \ + ((v128_t)__builtin_wasm_shuffle_i8x16( \ (__i8x16)(__a), (__i8x16)(__b), __c0, __c1, __c2, __c3, __c4, __c5, \ __c6, __c7, __c8, __c9, __c10, __c11, __c12, __c13, __c14, __c15)) -#define wasm_v16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ +#define wasm_i16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ __c7) \ - ((v128_t)__builtin_wasm_shuffle_v8x16( \ + ((v128_t)__builtin_wasm_shuffle_i8x16( \ (__i8x16)(__a), (__i8x16)(__b), (__c0)*2, (__c0)*2 + 1, (__c1)*2, \ (__c1)*2 + 1, (__c2)*2, (__c2)*2 + 1, (__c3)*2, (__c3)*2 + 1, (__c4)*2, \ (__c4)*2 + 1, (__c5)*2, (__c5)*2 + 1, (__c6)*2, (__c6)*2 + 1, (__c7)*2, \ (__c7)*2 + 1)) -#define wasm_v32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \ - ((v128_t)__builtin_wasm_shuffle_v8x16( \ +#define wasm_i32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \ + ((v128_t)__builtin_wasm_shuffle_i8x16( \ (__i8x16)(__a), (__i8x16)(__b), (__c0)*4, (__c0)*4 + 1, (__c0)*4 + 2, \ (__c0)*4 + 3, (__c1)*4, (__c1)*4 + 1, (__c1)*4 + 2, (__c1)*4 + 3, \ (__c2)*4, (__c2)*4 + 1, (__c2)*4 + 2, (__c2)*4 + 3, (__c3)*4, \ (__c3)*4 + 1, (__c3)*4 + 2, (__c3)*4 + 3)) -#define wasm_v64x2_shuffle(__a, __b, __c0, __c1) \ - ((v128_t)__builtin_wasm_shuffle_v8x16( \ +#define wasm_i64x2_shuffle(__a, __b, __c0, __c1) \ + ((v128_t)__builtin_wasm_shuffle_i8x16( \ (__i8x16)(__a), (__i8x16)(__b), (__c0)*8, (__c0)*8 + 1, (__c0)*8 + 2, \ (__c0)*8 + 3, (__c0)*8 + 4, (__c0)*8 + 5, (__c0)*8 + 6, (__c0)*8 + 7, \ (__c1)*8, (__c1)*8 + 1, (__c1)*8 + 2, (__c1)*8 + 3, (__c1)*8 + 4, \ (__c1)*8 + 5, (__c1)*8 + 6, (__c1)*8 + 7)) -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v8x16_swizzle(v128_t __a, +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_swizzle(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_swizzle_v8x16((__i8x16)__a, (__i8x16)__b); + return (v128_t)__builtin_wasm_swizzle_i8x16((__i8x16)__a, (__i8x16)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS @@ -1077,8 +1325,8 @@ wasm_i8x16_narrow_i16x8(v128_t __a, v128_t __b) { static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_narrow_i16x8(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_narrow_u_i8x16_i16x8((__u16x8)__a, - (__u16x8)__b); + return (v128_t)__builtin_wasm_narrow_u_i8x16_i16x8((__i16x8)__a, + (__i16x8)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS @@ -1089,12 +1337,12 @@ wasm_i16x8_narrow_i32x4(v128_t __a, v128_t __b) { static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_narrow_i32x4(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_narrow_u_i16x8_i32x4((__u32x4)__a, - (__u32x4)__b); + return (v128_t)__builtin_wasm_narrow_u_i16x8_i32x4((__i32x4)__a, + (__i32x4)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_widen_low_i8x16(v128_t __a) { +wasm_i16x8_extend_low_i8x16(v128_t __a) { return (v128_t) __builtin_convertvector( (__i8x8){((__i8x16)__a)[0], ((__i8x16)__a)[1], ((__i8x16)__a)[2], ((__i8x16)__a)[3], ((__i8x16)__a)[4], ((__i8x16)__a)[5], @@ -1103,7 +1351,7 @@ wasm_i16x8_widen_low_i8x16(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_widen_high_i8x16(v128_t __a) { +wasm_i16x8_extend_high_i8x16(v128_t __a) { return (v128_t) __builtin_convertvector( (__i8x8){((__i8x16)__a)[8], ((__i8x16)__a)[9], ((__i8x16)__a)[10], ((__i8x16)__a)[11], ((__i8x16)__a)[12], ((__i8x16)__a)[13], @@ -1112,7 +1360,7 @@ wasm_i16x8_widen_high_i8x16(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_widen_low_u8x16(v128_t __a) { +wasm_u16x8_extend_low_u8x16(v128_t __a) { return (v128_t) __builtin_convertvector( (__u8x8){((__u8x16)__a)[0], ((__u8x16)__a)[1], ((__u8x16)__a)[2], ((__u8x16)__a)[3], ((__u8x16)__a)[4], ((__u8x16)__a)[5], @@ -1121,7 +1369,7 @@ wasm_i16x8_widen_low_u8x16(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_widen_high_u8x16(v128_t __a) { +wasm_u16x8_extend_high_u8x16(v128_t __a) { return (v128_t) __builtin_convertvector( (__u8x8){((__u8x16)__a)[8], ((__u8x16)__a)[9], ((__u8x16)__a)[10], ((__u8x16)__a)[11], ((__u8x16)__a)[12], ((__u8x16)__a)[13], @@ -1130,7 +1378,7 @@ wasm_i16x8_widen_high_u8x16(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i32x4_widen_low_i16x8(v128_t __a) { +wasm_i32x4_extend_low_i16x8(v128_t __a) { return (v128_t) __builtin_convertvector( (__i16x4){((__i16x8)__a)[0], ((__i16x8)__a)[1], ((__i16x8)__a)[2], ((__i16x8)__a)[3]}, @@ -1138,7 +1386,7 @@ wasm_i32x4_widen_low_i16x8(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i32x4_widen_high_i16x8(v128_t __a) { +wasm_i32x4_extend_high_i16x8(v128_t __a) { return (v128_t) __builtin_convertvector( (__i16x4){((__i16x8)__a)[4], ((__i16x8)__a)[5], ((__i16x8)__a)[6], ((__i16x8)__a)[7]}, @@ -1146,7 +1394,7 @@ wasm_i32x4_widen_high_i16x8(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i32x4_widen_low_u16x8(v128_t __a) { +wasm_u32x4_extend_low_u16x8(v128_t __a) { return (v128_t) __builtin_convertvector( (__u16x4){((__u16x8)__a)[0], ((__u16x8)__a)[1], ((__u16x8)__a)[2], ((__u16x8)__a)[3]}, @@ -1154,14 +1402,333 @@ wasm_i32x4_widen_low_u16x8(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i32x4_widen_high_u16x8(v128_t __a) { +wasm_u32x4_extend_high_u16x8(v128_t __a) { return (v128_t) __builtin_convertvector( (__u16x4){((__u16x8)__a)[4], ((__u16x8)__a)[5], ((__u16x8)__a)[6], ((__u16x8)__a)[7]}, __u32x4); } +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i64x2_extend_low_i32x4(v128_t __a) { + return (v128_t) __builtin_convertvector( + (__i32x2){((__i32x4)__a)[0], ((__i32x4)__a)[1]}, __i64x2); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i64x2_extend_high_i32x4(v128_t __a) { + return (v128_t) __builtin_convertvector( + (__i32x2){((__i32x4)__a)[2], ((__i32x4)__a)[3]}, __i64x2); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u64x2_extend_low_u32x4(v128_t __a) { + return (v128_t) __builtin_convertvector( + (__u32x2){((__u32x4)__a)[0], ((__u32x4)__a)[1]}, __u64x2); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u64x2_extend_high_u32x4(v128_t __a) { + return (v128_t) __builtin_convertvector( + (__u32x2){((__u32x4)__a)[2], ((__u32x4)__a)[3]}, __u64x2); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i16x8_extadd_pairwise_i8x16(v128_t __a) { + return (v128_t)__builtin_wasm_extadd_pairwise_i8x16_s_i16x8((__i8x16)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u16x8_extadd_pairwise_u8x16(v128_t __a) { + return (v128_t)__builtin_wasm_extadd_pairwise_i8x16_u_i16x8((__u8x16)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i32x4_extadd_pairwise_i16x8(v128_t __a) { + return (v128_t)__builtin_wasm_extadd_pairwise_i16x8_s_i32x4((__i16x8)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u32x4_extadd_pairwise_u16x8(v128_t __a) { + return (v128_t)__builtin_wasm_extadd_pairwise_i16x8_u_i32x4((__u16x8)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i16x8_extmul_low_i8x16(v128_t __a, v128_t __b) { + return (v128_t)((__i16x8)wasm_i16x8_extend_low_i8x16(__a) * + (__i16x8)wasm_i16x8_extend_low_i8x16(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i16x8_extmul_high_i8x16(v128_t __a, v128_t __b) { + return (v128_t)((__i16x8)wasm_i16x8_extend_high_i8x16(__a) * + (__i16x8)wasm_i16x8_extend_high_i8x16(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u16x8_extmul_low_u8x16(v128_t __a, v128_t __b) { + return (v128_t)((__u16x8)wasm_u16x8_extend_low_u8x16(__a) * + (__u16x8)wasm_u16x8_extend_low_u8x16(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u16x8_extmul_high_u8x16(v128_t __a, v128_t __b) { + return (v128_t)((__u16x8)wasm_u16x8_extend_high_u8x16(__a) * + (__u16x8)wasm_u16x8_extend_high_u8x16(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i32x4_extmul_low_i16x8(v128_t __a, v128_t __b) { + return (v128_t)((__i32x4)wasm_i32x4_extend_low_i16x8(__a) * + (__i32x4)wasm_i32x4_extend_low_i16x8(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i32x4_extmul_high_i16x8(v128_t __a, v128_t __b) { + return (v128_t)((__i32x4)wasm_i32x4_extend_high_i16x8(__a) * + (__i32x4)wasm_i32x4_extend_high_i16x8(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u32x4_extmul_low_u16x8(v128_t __a, v128_t __b) { + return (v128_t)((__u32x4)wasm_u32x4_extend_low_u16x8(__a) * + (__u32x4)wasm_u32x4_extend_low_u16x8(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u32x4_extmul_high_u16x8(v128_t __a, v128_t __b) { + return (v128_t)((__u32x4)wasm_u32x4_extend_high_u16x8(__a) * + (__u32x4)wasm_u32x4_extend_high_u16x8(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i64x2_extmul_low_i32x4(v128_t __a, v128_t __b) { + return (v128_t)((__i64x2)wasm_i64x2_extend_low_i32x4(__a) * + (__i64x2)wasm_i64x2_extend_low_i32x4(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i64x2_extmul_high_i32x4(v128_t __a, v128_t __b) { + return (v128_t)((__i64x2)wasm_i64x2_extend_high_i32x4(__a) * + (__i64x2)wasm_i64x2_extend_high_i32x4(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u64x2_extmul_low_u32x4(v128_t __a, v128_t __b) { + return (v128_t)((__u64x2)wasm_u64x2_extend_low_u32x4(__a) * + (__u64x2)wasm_u64x2_extend_low_u32x4(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u64x2_extmul_high_u32x4(v128_t __a, v128_t __b) { + return (v128_t)((__u64x2)wasm_u64x2_extend_high_u32x4(__a) * + (__u64x2)wasm_u64x2_extend_high_u32x4(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_q15mulr_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_q15mulr_sat_s_i16x8((__i16x8)__a, (__i16x8)__b); +} + +// Old intrinsic names supported to ease transitioning to the standard names. Do +// not use these; they will be removed in the near future. + +#define __DEPRECATED_FN_ATTRS(__replacement) \ + __DEFAULT_FN_ATTRS __attribute__( \ + (deprecated("use " __replacement " instead", __replacement))) + +#define __WASM_STR(X) #X + +#ifdef __DEPRECATED +#define __DEPRECATED_WASM_MACRO(__name, __replacement) \ + _Pragma(__WASM_STR(GCC warning( \ + "'" __name "' is deprecated: use '" __replacement "' instead"))) +#else +#define __DEPRECATED_WASM_MACRO(__name, __replacement) +#endif + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_v128_load8_splat") +wasm_v8x16_load_splat(const void *__mem) { + return wasm_v128_load8_splat(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_v128_load16_splat") +wasm_v16x8_load_splat(const void *__mem) { + return wasm_v128_load16_splat(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_v128_load32_splat") +wasm_v32x4_load_splat(const void *__mem) { + return wasm_v128_load32_splat(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_v128_load64_splat") +wasm_v64x2_load_splat(const void *__mem) { + return wasm_v128_load64_splat(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i16x8_load8x8") +wasm_i16x8_load_8x8(const void *__mem) { + return wasm_i16x8_load8x8(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u16x8_load8x8") +wasm_u16x8_load_8x8(const void *__mem) { + return wasm_u16x8_load8x8(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i32x4_load16x4") +wasm_i32x4_load_16x4(const void *__mem) { + return wasm_i32x4_load16x4(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u32x4_load16x4") +wasm_u32x4_load_16x4(const void *__mem) { + return wasm_u32x4_load16x4(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i64x2_load32x2") +wasm_i64x2_load_32x2(const void *__mem) { + return wasm_i64x2_load32x2(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u64x2_load32x2") +wasm_u64x2_load_32x2(const void *__mem) { + return wasm_u64x2_load32x2(__mem); +} + +#define wasm_v8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ + __c7, __c8, __c9, __c10, __c11, __c12, __c13, \ + __c14, __c15) \ + __DEPRECATED_WASM_MACRO("wasm_v8x16_shuffle", "wasm_i8x16_shuffle") \ + wasm_i8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7, \ + __c8, __c9, __c10, __c11, __c12, __c13, __c14, __c15) + +#define wasm_v16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ + __c7) \ + __DEPRECATED_WASM_MACRO("wasm_v16x8_shuffle", "wasm_i16x8_shuffle") \ + wasm_i16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7) + +#define wasm_v32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \ + __DEPRECATED_WASM_MACRO("wasm_v32x4_shuffle", "wasm_i32x4_shuffle") \ + wasm_i32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) + +#define wasm_v64x2_shuffle(__a, __b, __c0, __c1) \ + __DEPRECATED_WASM_MACRO("wasm_v64x2_shuffle", "wasm_i64x2_shuffle") \ + wasm_i64x2_shuffle(__a, __b, __c0, __c1) + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i8x16_swizzle") +wasm_v8x16_swizzle(v128_t __a, v128_t __b) { + return wasm_i8x16_swizzle(__a, __b); +} + +static __inline__ bool __DEPRECATED_FN_ATTRS("wasm_v128_any_true") +wasm_i8x16_any_true(v128_t __a) { + return wasm_v128_any_true(__a); +} + +static __inline__ bool __DEPRECATED_FN_ATTRS("wasm_v128_any_true") +wasm_i16x8_any_true(v128_t __a) { + return wasm_v128_any_true(__a); +} + +static __inline__ bool __DEPRECATED_FN_ATTRS("wasm_v128_any_true") +wasm_i32x4_any_true(v128_t __a) { + return wasm_v128_any_true(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i8x16_add_sat") +wasm_i8x16_add_saturate(v128_t __a, v128_t __b) { + return wasm_i8x16_add_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u8x16_add_sat") +wasm_u8x16_add_saturate(v128_t __a, v128_t __b) { + return wasm_u8x16_add_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i8x16_sub_sat") +wasm_i8x16_sub_saturate(v128_t __a, v128_t __b) { + return wasm_i8x16_sub_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u8x16_sub_sat") +wasm_u8x16_sub_saturate(v128_t __a, v128_t __b) { + return wasm_u8x16_sub_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i16x8_add_sat") +wasm_i16x8_add_saturate(v128_t __a, v128_t __b) { + return wasm_i16x8_add_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u16x8_add_sat") +wasm_u16x8_add_saturate(v128_t __a, v128_t __b) { + return wasm_u16x8_add_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i16x8_sub_sat") +wasm_i16x8_sub_saturate(v128_t __a, v128_t __b) { + return wasm_i16x8_sub_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u16x8_sub_sat") +wasm_u16x8_sub_saturate(v128_t __a, v128_t __b) { + return wasm_u16x8_sub_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i16x8_extend_low_i8x16") +wasm_i16x8_widen_low_i8x16(v128_t __a) { + return wasm_i16x8_extend_low_i8x16(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i16x8_extend_high_i8x16") +wasm_i16x8_widen_high_i8x16(v128_t __a) { + return wasm_i16x8_extend_high_i8x16(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u16x8_extend_low_u8x16") +wasm_i16x8_widen_low_u8x16(v128_t __a) { + return wasm_u16x8_extend_low_u8x16(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u16x8_extend_high_u8x16") +wasm_i16x8_widen_high_u8x16(v128_t __a) { + return wasm_u16x8_extend_high_u8x16(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i32x4_extend_low_i16x8") +wasm_i32x4_widen_low_i16x8(v128_t __a) { + return wasm_i32x4_extend_low_i16x8(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i32x4_extend_high_i16x8") +wasm_i32x4_widen_high_i16x8(v128_t __a) { + return wasm_i32x4_extend_high_i16x8(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u32x4_extend_low_u16x8") +wasm_i32x4_widen_low_u16x8(v128_t __a) { + return wasm_u32x4_extend_low_u16x8(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u32x4_extend_high_u16x8") +wasm_i32x4_widen_high_u16x8(v128_t __a) { + return wasm_u32x4_extend_high_u16x8(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i32x4_trunc_sat_f32x4") +wasm_i32x4_trunc_saturate_f32x4(v128_t __a) { + return wasm_i32x4_trunc_sat_f32x4(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u32x4_trunc_sat_f32x4") +wasm_u32x4_trunc_saturate_f32x4(v128_t __a) { + return wasm_u32x4_trunc_sat_f32x4(__a); +} + // Undefine helper macros #undef __DEFAULT_FN_ATTRS +#undef __DEPRECATED_FN_ATTRS #endif // __WASM_SIMD128_H